var/home/core/zuul-output/0000755000175000017500000000000015071276556014543 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071313235015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005222652215071313227017704 0ustar rootrootOct 07 20:55:48 crc systemd[1]: Starting Kubernetes Kubelet... Oct 07 20:55:48 crc restorecon[4658]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:48 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 20:55:49 crc restorecon[4658]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 07 20:55:50 crc kubenswrapper[4926]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 20:55:50 crc kubenswrapper[4926]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 07 20:55:50 crc kubenswrapper[4926]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 20:55:50 crc kubenswrapper[4926]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 20:55:50 crc kubenswrapper[4926]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 07 20:55:50 crc kubenswrapper[4926]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.391190 4926 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402189 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402247 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402257 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402266 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402275 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402286 4926 feature_gate.go:330] unrecognized feature gate: Example Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402295 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402303 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402312 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402319 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402327 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402335 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402343 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402364 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402372 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402380 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402388 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402396 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402403 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402411 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402418 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402426 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402433 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402441 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402448 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402459 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402471 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402481 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402490 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402498 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402506 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402514 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402523 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402531 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402538 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402548 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402558 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402566 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402574 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402582 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402590 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402599 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402610 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402620 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402629 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402637 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402645 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402653 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402665 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402674 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402684 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402692 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402702 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402709 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402717 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402725 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402733 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402746 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402753 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402762 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402769 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402777 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402785 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402792 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402800 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402807 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402816 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402824 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402834 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402841 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.402849 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404067 4926 flags.go:64] FLAG: --address="0.0.0.0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404089 4926 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404106 4926 flags.go:64] FLAG: --anonymous-auth="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404117 4926 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404129 4926 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404138 4926 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404149 4926 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404161 4926 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404170 4926 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404179 4926 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404190 4926 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404224 4926 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404235 4926 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404245 4926 flags.go:64] FLAG: --cgroup-root="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404254 4926 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404263 4926 flags.go:64] FLAG: --client-ca-file="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404272 4926 flags.go:64] FLAG: --cloud-config="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404281 4926 flags.go:64] FLAG: --cloud-provider="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404291 4926 flags.go:64] FLAG: --cluster-dns="[]" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404302 4926 flags.go:64] FLAG: --cluster-domain="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404311 4926 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404320 4926 flags.go:64] FLAG: --config-dir="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404329 4926 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404339 4926 flags.go:64] FLAG: --container-log-max-files="5" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404350 4926 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404359 4926 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404369 4926 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404378 4926 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404387 4926 flags.go:64] FLAG: --contention-profiling="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404396 4926 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404405 4926 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404416 4926 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404425 4926 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404437 4926 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404446 4926 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404455 4926 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404464 4926 flags.go:64] FLAG: --enable-load-reader="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404473 4926 flags.go:64] FLAG: --enable-server="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404483 4926 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404495 4926 flags.go:64] FLAG: --event-burst="100" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404504 4926 flags.go:64] FLAG: --event-qps="50" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404513 4926 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404522 4926 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404531 4926 flags.go:64] FLAG: --eviction-hard="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404542 4926 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404551 4926 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404561 4926 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404570 4926 flags.go:64] FLAG: --eviction-soft="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404579 4926 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404588 4926 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404597 4926 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404606 4926 flags.go:64] FLAG: --experimental-mounter-path="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404614 4926 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404623 4926 flags.go:64] FLAG: --fail-swap-on="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404632 4926 flags.go:64] FLAG: --feature-gates="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404642 4926 flags.go:64] FLAG: --file-check-frequency="20s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404651 4926 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404660 4926 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404670 4926 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404679 4926 flags.go:64] FLAG: --healthz-port="10248" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404688 4926 flags.go:64] FLAG: --help="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404697 4926 flags.go:64] FLAG: --hostname-override="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404706 4926 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404715 4926 flags.go:64] FLAG: --http-check-frequency="20s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404724 4926 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404733 4926 flags.go:64] FLAG: --image-credential-provider-config="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404741 4926 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404751 4926 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404760 4926 flags.go:64] FLAG: --image-service-endpoint="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404769 4926 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404778 4926 flags.go:64] FLAG: --kube-api-burst="100" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404787 4926 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404797 4926 flags.go:64] FLAG: --kube-api-qps="50" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404805 4926 flags.go:64] FLAG: --kube-reserved="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404814 4926 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404823 4926 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404833 4926 flags.go:64] FLAG: --kubelet-cgroups="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404843 4926 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404851 4926 flags.go:64] FLAG: --lock-file="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404860 4926 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404869 4926 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404879 4926 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404892 4926 flags.go:64] FLAG: --log-json-split-stream="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404900 4926 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404909 4926 flags.go:64] FLAG: --log-text-split-stream="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404918 4926 flags.go:64] FLAG: --logging-format="text" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404927 4926 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404937 4926 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404945 4926 flags.go:64] FLAG: --manifest-url="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404954 4926 flags.go:64] FLAG: --manifest-url-header="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404965 4926 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404975 4926 flags.go:64] FLAG: --max-open-files="1000000" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404986 4926 flags.go:64] FLAG: --max-pods="110" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.404995 4926 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405004 4926 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405013 4926 flags.go:64] FLAG: --memory-manager-policy="None" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405022 4926 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405031 4926 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405040 4926 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405050 4926 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405069 4926 flags.go:64] FLAG: --node-status-max-images="50" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405078 4926 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405087 4926 flags.go:64] FLAG: --oom-score-adj="-999" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405097 4926 flags.go:64] FLAG: --pod-cidr="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405107 4926 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405120 4926 flags.go:64] FLAG: --pod-manifest-path="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405129 4926 flags.go:64] FLAG: --pod-max-pids="-1" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405138 4926 flags.go:64] FLAG: --pods-per-core="0" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405147 4926 flags.go:64] FLAG: --port="10250" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405156 4926 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405165 4926 flags.go:64] FLAG: --provider-id="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405174 4926 flags.go:64] FLAG: --qos-reserved="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405183 4926 flags.go:64] FLAG: --read-only-port="10255" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405214 4926 flags.go:64] FLAG: --register-node="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405224 4926 flags.go:64] FLAG: --register-schedulable="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405232 4926 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405247 4926 flags.go:64] FLAG: --registry-burst="10" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405255 4926 flags.go:64] FLAG: --registry-qps="5" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405264 4926 flags.go:64] FLAG: --reserved-cpus="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405273 4926 flags.go:64] FLAG: --reserved-memory="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405284 4926 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405293 4926 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405302 4926 flags.go:64] FLAG: --rotate-certificates="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405311 4926 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405320 4926 flags.go:64] FLAG: --runonce="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405328 4926 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405337 4926 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405347 4926 flags.go:64] FLAG: --seccomp-default="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405355 4926 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405364 4926 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405373 4926 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405382 4926 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405392 4926 flags.go:64] FLAG: --storage-driver-password="root" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405400 4926 flags.go:64] FLAG: --storage-driver-secure="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405409 4926 flags.go:64] FLAG: --storage-driver-table="stats" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405418 4926 flags.go:64] FLAG: --storage-driver-user="root" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405427 4926 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405436 4926 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405445 4926 flags.go:64] FLAG: --system-cgroups="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405455 4926 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405468 4926 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405477 4926 flags.go:64] FLAG: --tls-cert-file="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405487 4926 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405497 4926 flags.go:64] FLAG: --tls-min-version="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405520 4926 flags.go:64] FLAG: --tls-private-key-file="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405529 4926 flags.go:64] FLAG: --topology-manager-policy="none" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405538 4926 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405546 4926 flags.go:64] FLAG: --topology-manager-scope="container" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405556 4926 flags.go:64] FLAG: --v="2" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405566 4926 flags.go:64] FLAG: --version="false" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405577 4926 flags.go:64] FLAG: --vmodule="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405587 4926 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.405597 4926 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405837 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405850 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405861 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405870 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405879 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405888 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405897 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405906 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405914 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405922 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405930 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405938 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405945 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405953 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405960 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405968 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405976 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405983 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405991 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.405998 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406006 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406022 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406031 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406041 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406049 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406056 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406065 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406072 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406080 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406087 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406095 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406104 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406112 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406119 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406127 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406137 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406147 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406156 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406164 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406173 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406181 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406189 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406221 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406229 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406239 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406248 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406256 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406264 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406271 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406279 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406287 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406295 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406303 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406313 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406321 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406331 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406339 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406347 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406356 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406364 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406372 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406379 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406387 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406395 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406403 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406411 4926 feature_gate.go:330] unrecognized feature gate: Example Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406418 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406426 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406434 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406442 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.406450 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.406471 4926 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.422763 4926 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.422808 4926 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.422943 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.422957 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.422967 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.422976 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.422985 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.422993 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423000 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423008 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423016 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423024 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423032 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423042 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423053 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423064 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423074 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423083 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423092 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423099 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423107 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423115 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423123 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423131 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423139 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423147 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423154 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423162 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423170 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423178 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423185 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423216 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423224 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423232 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423242 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423334 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423343 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423354 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423362 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423371 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423379 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423388 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423397 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423406 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423415 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423424 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423432 4926 feature_gate.go:330] unrecognized feature gate: Example Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423441 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423449 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423460 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423471 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423480 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423490 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423499 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423508 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423516 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423524 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423532 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423540 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423548 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423557 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423565 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423573 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423581 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423589 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423596 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423604 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423611 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423619 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423628 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423636 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423646 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423655 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.423668 4926 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423922 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423936 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423947 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423955 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423964 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423972 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423980 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.423990 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424000 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424008 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424017 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424024 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424032 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424042 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424052 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424062 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424070 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424078 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424086 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424095 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424105 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424113 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424121 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424128 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424136 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424144 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424151 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424159 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424167 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424174 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424183 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424217 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424226 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424235 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424243 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424252 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424260 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424268 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424277 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424285 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424293 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424301 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424309 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424316 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424324 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424331 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424342 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424351 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424359 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424367 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424375 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424383 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424391 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424398 4926 feature_gate.go:330] unrecognized feature gate: Example Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424406 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424414 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424422 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424430 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424438 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424446 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424453 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424461 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424469 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424477 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424485 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424492 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424500 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424509 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424516 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424524 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.424532 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.424544 4926 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.425835 4926 server.go:940] "Client rotation is on, will bootstrap in background" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.431819 4926 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.431950 4926 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.434366 4926 server.go:997] "Starting client certificate rotation" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.434415 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.437785 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-25 02:55:45.016761309 +0000 UTC Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.437911 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1877h59m54.578855984s for next certificate rotation Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.462072 4926 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.470012 4926 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.491962 4926 log.go:25] "Validated CRI v1 runtime API" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.532168 4926 log.go:25] "Validated CRI v1 image API" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.535021 4926 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.541292 4926 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-07-20-50-59-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.541337 4926 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.570933 4926 manager.go:217] Machine: {Timestamp:2025-10-07 20:55:50.566003221 +0000 UTC m=+0.604584451 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3 BootID:86d54cec-6d51-4e27-bc1f-9a644fe10e6e Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:df:21:04 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:df:21:04 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ee:9a:c9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1a:c7:0b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:63:2f:ef Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:bb:55:d3 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1e:80:df:7f:ae:74 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3e:e1:c6:a6:eb:60 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.571408 4926 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.571742 4926 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.573295 4926 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.573625 4926 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.573692 4926 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.576188 4926 topology_manager.go:138] "Creating topology manager with none policy" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.576265 4926 container_manager_linux.go:303] "Creating device plugin manager" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.576906 4926 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.576942 4926 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.577925 4926 state_mem.go:36] "Initialized new in-memory state store" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.578113 4926 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.582148 4926 kubelet.go:418] "Attempting to sync node with API server" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.582222 4926 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.582263 4926 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.582292 4926 kubelet.go:324] "Adding apiserver pod source" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.582316 4926 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.588751 4926 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.589887 4926 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.592083 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.592095 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.592345 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.592401 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.592760 4926 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595649 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595703 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595723 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595742 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595772 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595791 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595811 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595840 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595861 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595879 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595906 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.595926 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.596736 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.597508 4926 server.go:1280] "Started kubelet" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.597575 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.597699 4926 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.597726 4926 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.598569 4926 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 07 20:55:50 crc systemd[1]: Started Kubernetes Kubelet. Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.603919 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.604011 4926 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.604128 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 21:29:47.847698644 +0000 UTC Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.604161 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1152h33m57.24354033s for next certificate rotation Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.604261 4926 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.604297 4926 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.604317 4926 server.go:460] "Adding debug handlers to kubelet server" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.605363 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.605968 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.605990 4926 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.606327 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.606351 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.58:6443: connect: connection refused" interval="200ms" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.606925 4926 factory.go:55] Registering systemd factory Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.606966 4926 factory.go:221] Registration of the systemd container factory successfully Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.607473 4926 factory.go:153] Registering CRI-O factory Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.607493 4926 factory.go:221] Registration of the crio container factory successfully Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.607577 4926 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.607601 4926 factory.go:103] Registering Raw factory Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.607617 4926 manager.go:1196] Started watching for new ooms in manager Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.613805 4926 manager.go:319] Starting recovery of all containers Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.611422 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.58:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c50f10b088061 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 20:55:50.597468257 +0000 UTC m=+0.636049477,LastTimestamp:2025-10-07 20:55:50.597468257 +0000 UTC m=+0.636049477,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629270 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629376 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629399 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629416 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629435 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629455 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629471 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629489 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629509 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629527 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629546 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629564 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629582 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629605 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629625 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629644 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629672 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629716 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629736 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629754 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629774 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629792 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629811 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629828 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629864 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629882 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629904 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629924 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629943 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629961 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629979 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.629996 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630080 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630099 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630116 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630133 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630151 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630169 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630186 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630237 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630289 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630315 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630333 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630353 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630370 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630425 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630445 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630462 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630542 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630561 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630580 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630598 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630625 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630644 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630663 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630682 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630738 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630756 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630776 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630795 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630842 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630860 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630879 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630896 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630917 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630935 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630953 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630971 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.630996 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631013 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631031 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631057 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631111 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631128 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631145 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631170 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631188 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631238 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631264 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631282 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631314 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631333 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631399 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631418 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631435 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631453 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631472 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631489 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631515 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631533 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631558 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631591 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631625 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631650 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631674 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631701 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631729 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631753 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631779 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631802 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631829 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631853 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631880 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631909 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631942 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631969 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.631997 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632020 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632039 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632058 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632076 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632093 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632159 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632179 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632231 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632260 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632278 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632296 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632313 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632373 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632396 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632420 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632445 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632467 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632489 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632512 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632536 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632559 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632586 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632609 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632634 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632658 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632681 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632704 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632730 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632756 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632778 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632802 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632825 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632848 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632871 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632893 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632919 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632941 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632969 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.632992 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633016 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633040 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633062 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633086 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633108 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633130 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633153 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633176 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633234 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633257 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633279 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633301 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633326 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633352 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633380 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633404 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633429 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633453 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633478 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633503 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633528 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633558 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633584 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633611 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633636 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633661 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633686 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633713 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633737 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633761 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633787 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633810 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633834 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633859 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633884 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633912 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633936 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633963 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.633988 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634041 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634086 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634121 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634155 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634178 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634235 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634264 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634299 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634325 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634349 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634375 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.634400 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.637995 4926 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638066 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638103 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638128 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638152 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638177 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638233 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638257 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638286 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638315 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638341 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638364 4926 reconstruct.go:97] "Volume reconstruction finished" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.638379 4926 reconciler.go:26] "Reconciler: start to sync state" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.649456 4926 manager.go:324] Recovery completed Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.665026 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.667004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.667060 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.667073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.668120 4926 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.668161 4926 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.668183 4926 state_mem.go:36] "Initialized new in-memory state store" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.674457 4926 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.677517 4926 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.677564 4926 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.677595 4926 kubelet.go:2335] "Starting kubelet main sync loop" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.677650 4926 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.680504 4926 policy_none.go:49] "None policy: Start" Oct 07 20:55:50 crc kubenswrapper[4926]: W1007 20:55:50.681214 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.681565 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.681492 4926 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.681881 4926 state_mem.go:35] "Initializing new in-memory state store" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.706158 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.748745 4926 manager.go:334] "Starting Device Plugin manager" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.748823 4926 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.748840 4926 server.go:79] "Starting device plugin registration server" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.749399 4926 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.749417 4926 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.749578 4926 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.749682 4926 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.749691 4926 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.760384 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.778765 4926 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.778928 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.780574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.780635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.780660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.780867 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.781167 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.781264 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.783479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.783548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.783632 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.783922 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.784980 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.785532 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.785582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.785852 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.785578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.785694 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.786011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.786160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.787419 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.787609 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.787658 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.788545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.789027 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.789107 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.789156 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.789661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.789726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790379 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.790794 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.791926 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.793171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.793246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.793266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.807269 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.58:6443: connect: connection refused" interval="400ms" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.840414 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.840494 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.840584 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.840708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.840838 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.840958 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841021 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841068 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841119 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841164 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841251 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841348 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841389 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.841454 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.849558 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.851036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.851120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.851146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.851220 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:55:50 crc kubenswrapper[4926]: E1007 20:55:50.851939 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.58:6443: connect: connection refused" node="crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.943383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.943676 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.943918 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.944101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.944256 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.943495 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.943744 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.943960 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.944512 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.944675 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.944838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.944976 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945170 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945466 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945633 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945540 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945250 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945721 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945800 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945927 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945963 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.945996 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946015 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946028 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946045 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946059 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946075 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946139 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:50 crc kubenswrapper[4926]: I1007 20:55:50.946549 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.052256 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.053536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.053568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.053579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.053608 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:55:51 crc kubenswrapper[4926]: E1007 20:55:51.053893 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.58:6443: connect: connection refused" node="crc" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.132016 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.147338 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.175236 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.183781 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-760120437db87f5639f3e6c50588d313f1498bdd4eeb17dc5af92c756852e6a3 WatchSource:0}: Error finding container 760120437db87f5639f3e6c50588d313f1498bdd4eeb17dc5af92c756852e6a3: Status 404 returned error can't find the container with id 760120437db87f5639f3e6c50588d313f1498bdd4eeb17dc5af92c756852e6a3 Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.188281 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.189726 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ac84934d151c7fca4033173a588c8d44e5679b4e7e763d6f8814b4232531ec60 WatchSource:0}: Error finding container ac84934d151c7fca4033173a588c8d44e5679b4e7e763d6f8814b4232531ec60: Status 404 returned error can't find the container with id ac84934d151c7fca4033173a588c8d44e5679b4e7e763d6f8814b4232531ec60 Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.196425 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.200070 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ce2846864029244f30f184f0306781de0355c9582b724657b607c791cc4fb554 WatchSource:0}: Error finding container ce2846864029244f30f184f0306781de0355c9582b724657b607c791cc4fb554: Status 404 returned error can't find the container with id ce2846864029244f30f184f0306781de0355c9582b724657b607c791cc4fb554 Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.206823 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-889a3d47eb239d9d3cf4d173c3982eba42e463d9a43b66be80f9249a991e0d06 WatchSource:0}: Error finding container 889a3d47eb239d9d3cf4d173c3982eba42e463d9a43b66be80f9249a991e0d06: Status 404 returned error can't find the container with id 889a3d47eb239d9d3cf4d173c3982eba42e463d9a43b66be80f9249a991e0d06 Oct 07 20:55:51 crc kubenswrapper[4926]: E1007 20:55:51.208409 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.58:6443: connect: connection refused" interval="800ms" Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.219771 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-5ce2b2457a76d2acd6e0b116f05d3342341247bd75476376a397d070f9c53acb WatchSource:0}: Error finding container 5ce2b2457a76d2acd6e0b116f05d3342341247bd75476376a397d070f9c53acb: Status 404 returned error can't find the container with id 5ce2b2457a76d2acd6e0b116f05d3342341247bd75476376a397d070f9c53acb Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.454273 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.456899 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.456976 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.456997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.457049 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:55:51 crc kubenswrapper[4926]: E1007 20:55:51.457961 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.58:6443: connect: connection refused" node="crc" Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.484080 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:51 crc kubenswrapper[4926]: E1007 20:55:51.484208 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.599091 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.652259 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:51 crc kubenswrapper[4926]: E1007 20:55:51.652390 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.682813 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"760120437db87f5639f3e6c50588d313f1498bdd4eeb17dc5af92c756852e6a3"} Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.684843 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ac84934d151c7fca4033173a588c8d44e5679b4e7e763d6f8814b4232531ec60"} Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.686299 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5ce2b2457a76d2acd6e0b116f05d3342341247bd75476376a397d070f9c53acb"} Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.688053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"889a3d47eb239d9d3cf4d173c3982eba42e463d9a43b66be80f9249a991e0d06"} Oct 07 20:55:51 crc kubenswrapper[4926]: I1007 20:55:51.689295 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ce2846864029244f30f184f0306781de0355c9582b724657b607c791cc4fb554"} Oct 07 20:55:51 crc kubenswrapper[4926]: W1007 20:55:51.937230 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:51 crc kubenswrapper[4926]: E1007 20:55:51.937342 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:52 crc kubenswrapper[4926]: E1007 20:55:52.009741 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.58:6443: connect: connection refused" interval="1.6s" Oct 07 20:55:52 crc kubenswrapper[4926]: W1007 20:55:52.014618 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:52 crc kubenswrapper[4926]: E1007 20:55:52.014732 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.258288 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.259716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.259749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.259762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.259792 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:55:52 crc kubenswrapper[4926]: E1007 20:55:52.260535 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.58:6443: connect: connection refused" node="crc" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.598846 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.696651 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.696720 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.696753 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.699771 4926 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9" exitCode=0 Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.699825 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.699985 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.701562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.701608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.701622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.702437 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140" exitCode=0 Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.702603 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.702523 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.704894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.704931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.704945 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.707806 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.708677 4926 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3" exitCode=0 Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.708757 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.708784 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.709226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.709257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.709270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.709824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.710049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.710073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.711679 4926 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac" exitCode=0 Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.711761 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.711755 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac"} Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.713316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.713354 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:52 crc kubenswrapper[4926]: I1007 20:55:52.713365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:53 crc kubenswrapper[4926]: E1007 20:55:53.011015 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.58:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c50f10b088061 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 20:55:50.597468257 +0000 UTC m=+0.636049477,LastTimestamp:2025-10-07 20:55:50.597468257 +0000 UTC m=+0.636049477,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.598799 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:53 crc kubenswrapper[4926]: E1007 20:55:53.610406 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.58:6443: connect: connection refused" interval="3.2s" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.715987 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.716054 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.717855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.717888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.717898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.722045 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.721993 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.722118 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.722140 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.723166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.723210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.723221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.726894 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.727000 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.729832 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.729874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.729888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.732456 4926 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541" exitCode=0 Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.732561 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.732550 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.733623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.733663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.733676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.738906 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.738938 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.738953 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188"} Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.738966 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40"} Oct 07 20:55:53 crc kubenswrapper[4926]: W1007 20:55:53.829746 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:53 crc kubenswrapper[4926]: E1007 20:55:53.829878 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.861242 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.863071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.863157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.863172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:53 crc kubenswrapper[4926]: I1007 20:55:53.863232 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:55:53 crc kubenswrapper[4926]: E1007 20:55:53.863995 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.58:6443: connect: connection refused" node="crc" Oct 07 20:55:54 crc kubenswrapper[4926]: W1007 20:55:54.136705 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.58:6443: connect: connection refused Oct 07 20:55:54 crc kubenswrapper[4926]: E1007 20:55:54.137027 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.58:6443: connect: connection refused" logger="UnhandledError" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.745604 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7"} Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.745740 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.747107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.747168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.747187 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.748408 4926 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3" exitCode=0 Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.748439 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3"} Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.748568 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.748645 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.748670 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.749941 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750022 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750729 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750863 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.751002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.751179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.751252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.750808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.751401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:54 crc kubenswrapper[4926]: I1007 20:55:54.751431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.756948 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e"} Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.757031 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3"} Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.757057 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.757120 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.757123 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.757058 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11"} Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.759018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.759075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.759092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.759897 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.759931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.759949 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.868956 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.869156 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.870555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.870596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:55 crc kubenswrapper[4926]: I1007 20:55:55.870618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.103590 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.767768 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa"} Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.767852 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b"} Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.767940 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.767961 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.768005 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.769712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.769762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.769723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.769781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.769802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:56 crc kubenswrapper[4926]: I1007 20:55:56.769905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.064737 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.066274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.066322 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.066338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.066378 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.770980 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.772319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.772386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.772410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.856420 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.856553 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.856593 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.857818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.857874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:57 crc kubenswrapper[4926]: I1007 20:55:57.857897 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.052580 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.052880 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.054661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.054710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.054729 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.377355 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.377595 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.379069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.379121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:55:59 crc kubenswrapper[4926]: I1007 20:55:59.379139 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:00 crc kubenswrapper[4926]: E1007 20:56:00.760557 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 20:56:00 crc kubenswrapper[4926]: I1007 20:56:00.971601 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 07 20:56:00 crc kubenswrapper[4926]: I1007 20:56:00.971885 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:00 crc kubenswrapper[4926]: I1007 20:56:00.973385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:00 crc kubenswrapper[4926]: I1007 20:56:00.973434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:00 crc kubenswrapper[4926]: I1007 20:56:00.973451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.744727 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.744912 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.747274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.747329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.747352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.753427 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.782652 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.783941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.783992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.784011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:01 crc kubenswrapper[4926]: I1007 20:56:01.790382 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:02 crc kubenswrapper[4926]: I1007 20:56:02.785138 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:02 crc kubenswrapper[4926]: I1007 20:56:02.786510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:02 crc kubenswrapper[4926]: I1007 20:56:02.786553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:02 crc kubenswrapper[4926]: I1007 20:56:02.786572 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:02 crc kubenswrapper[4926]: I1007 20:56:02.984882 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.435561 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.435799 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.437170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.437450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.437476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.788097 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.789739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.789803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:03 crc kubenswrapper[4926]: I1007 20:56:03.789821 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:04 crc kubenswrapper[4926]: I1007 20:56:04.600003 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 07 20:56:04 crc kubenswrapper[4926]: W1007 20:56:04.652850 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 07 20:56:04 crc kubenswrapper[4926]: I1007 20:56:04.652948 4926 trace.go:236] Trace[1655316752]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 20:55:54.651) (total time: 10001ms): Oct 07 20:56:04 crc kubenswrapper[4926]: Trace[1655316752]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:56:04.652) Oct 07 20:56:04 crc kubenswrapper[4926]: Trace[1655316752]: [10.001648648s] [10.001648648s] END Oct 07 20:56:04 crc kubenswrapper[4926]: E1007 20:56:04.652969 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 07 20:56:04 crc kubenswrapper[4926]: W1007 20:56:04.724187 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 07 20:56:04 crc kubenswrapper[4926]: I1007 20:56:04.724317 4926 trace.go:236] Trace[922897795]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 20:55:54.722) (total time: 10001ms): Oct 07 20:56:04 crc kubenswrapper[4926]: Trace[922897795]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:56:04.724) Oct 07 20:56:04 crc kubenswrapper[4926]: Trace[922897795]: [10.001559376s] [10.001559376s] END Oct 07 20:56:04 crc kubenswrapper[4926]: E1007 20:56:04.724342 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 07 20:56:05 crc kubenswrapper[4926]: I1007 20:56:05.534126 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 20:56:05 crc kubenswrapper[4926]: I1007 20:56:05.534276 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 20:56:05 crc kubenswrapper[4926]: I1007 20:56:05.543137 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 20:56:05 crc kubenswrapper[4926]: I1007 20:56:05.543271 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 20:56:05 crc kubenswrapper[4926]: I1007 20:56:05.985569 4926 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 20:56:05 crc kubenswrapper[4926]: I1007 20:56:05.985679 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 20:56:07 crc kubenswrapper[4926]: I1007 20:56:07.866731 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:56:07 crc kubenswrapper[4926]: I1007 20:56:07.867053 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:07 crc kubenswrapper[4926]: I1007 20:56:07.868807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:07 crc kubenswrapper[4926]: I1007 20:56:07.868867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:07 crc kubenswrapper[4926]: I1007 20:56:07.868884 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:07 crc kubenswrapper[4926]: I1007 20:56:07.872563 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:56:08 crc kubenswrapper[4926]: I1007 20:56:08.362817 4926 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 07 20:56:08 crc kubenswrapper[4926]: I1007 20:56:08.801819 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:08 crc kubenswrapper[4926]: I1007 20:56:08.802918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:08 crc kubenswrapper[4926]: I1007 20:56:08.802980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:08 crc kubenswrapper[4926]: I1007 20:56:08.803004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.018087 4926 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.506670 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.508443 4926 trace.go:236] Trace[1867245733]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 20:55:58.964) (total time: 11543ms): Oct 07 20:56:10 crc kubenswrapper[4926]: Trace[1867245733]: ---"Objects listed" error: 11543ms (20:56:10.508) Oct 07 20:56:10 crc kubenswrapper[4926]: Trace[1867245733]: [11.543619148s] [11.543619148s] END Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.508471 4926 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.511582 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.512270 4926 trace.go:236] Trace[727708365]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 20:55:59.322) (total time: 11189ms): Oct 07 20:56:10 crc kubenswrapper[4926]: Trace[727708365]: ---"Objects listed" error: 11189ms (20:56:10.512) Oct 07 20:56:10 crc kubenswrapper[4926]: Trace[727708365]: [11.189961495s] [11.189961495s] END Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.512295 4926 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.516088 4926 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.572630 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58606->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.572687 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58606->192.168.126.11:17697: read: connection reset by peer" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.572719 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58614->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.572810 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58614->192.168.126.11:17697: read: connection reset by peer" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.573120 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.573266 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.573709 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.573754 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.595850 4926 apiserver.go:52] "Watching apiserver" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.599800 4926 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.600063 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.600592 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.600855 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.600958 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.601013 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.601038 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.601088 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.601120 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.601168 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.601288 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.604108 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.604582 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.604794 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.604809 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.605907 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.606184 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.606397 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.606577 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.606718 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.607719 4926 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616758 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616796 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616816 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616837 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616853 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616869 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616886 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616976 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.616996 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617014 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617031 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617048 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617065 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617082 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617115 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617130 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617147 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617163 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617218 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617247 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617260 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617276 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617296 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617309 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617324 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617338 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617353 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617368 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617383 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617402 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617417 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617432 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617448 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617466 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617480 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617494 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617509 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617524 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617580 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617594 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617613 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617632 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617649 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617663 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617678 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617699 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617715 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617729 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617745 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617769 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617784 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617800 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617817 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617832 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617846 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617860 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617875 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617889 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617905 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617919 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617934 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617950 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617964 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617979 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.617994 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618012 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618030 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618044 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618060 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618088 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618103 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618121 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618136 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618150 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618165 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618180 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618227 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618243 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618257 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618273 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618287 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618302 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618317 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618335 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618349 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618364 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618380 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618394 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618409 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618423 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618438 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618452 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618468 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618483 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618498 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618513 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618527 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618542 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618559 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618577 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618593 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618610 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618626 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618642 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618658 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618673 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618688 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618703 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618721 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618738 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618769 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618785 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618799 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618815 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618831 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618847 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618864 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618897 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618913 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618929 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618945 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618961 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618977 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.618993 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619010 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619026 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619041 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619056 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619070 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619088 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619106 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619125 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619141 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619158 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619174 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619210 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619230 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619246 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619263 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619280 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619296 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619311 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619327 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619343 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619360 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619377 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619393 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619409 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619426 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619443 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619459 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619476 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619492 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619509 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619525 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619541 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619558 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619585 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619602 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619618 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619634 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619651 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619668 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619686 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619703 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619721 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619738 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619755 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619770 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619786 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619804 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619822 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619838 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619854 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619871 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619888 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619903 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619919 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619936 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619953 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619970 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.619986 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620005 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620021 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620036 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620051 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620069 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620085 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620139 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620159 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620179 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.620997 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621002 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621068 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621097 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621153 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621259 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621284 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621437 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621467 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621475 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621554 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621574 4926 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621813 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.621998 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.622065 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.622357 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.622514 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.622553 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.622863 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.623013 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.623075 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.624073 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.625700 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.625846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.626067 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:56:11.126036724 +0000 UTC m=+21.164617904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.625955 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.626352 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.627095 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.627085 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.627501 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.627875 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.628554 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.628747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629063 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629074 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629524 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629544 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629552 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629590 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.629985 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.630016 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.630104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.630931 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.631346 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.631362 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.631721 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.632021 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.632408 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.632864 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.632940 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.633171 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.633562 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.633724 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.633780 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634241 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634295 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634464 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634506 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634632 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.634710 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635462 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635552 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635569 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635697 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635704 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635741 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.635980 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636031 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636061 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636118 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636176 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636308 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636633 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.636691 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.637094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.637124 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.637272 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.637814 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.637853 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.638096 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.638689 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.638847 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.638909 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639352 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639520 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639748 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639690 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639812 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639822 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.640151 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.640148 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.640218 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:11.140203585 +0000 UTC m=+21.178784735 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.640262 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.640288 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:11.140282177 +0000 UTC m=+21.178863327 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.640709 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.640820 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.640842 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.641063 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.641254 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.640957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.641562 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.641620 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.641719 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.641802 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.642017 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.639875 4926 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.642143 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.643085 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.643232 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.643309 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.643451 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.643628 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.643920 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.644137 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.644169 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.644457 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.644455 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645006 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645180 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645554 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645650 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645848 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.645994 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646030 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.644857 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646321 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646354 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646495 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646556 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646711 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646919 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646953 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.646994 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.647104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.649847 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.650183 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.650611 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.653643 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.653741 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.654122 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.658301 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.658648 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.658850 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.658870 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.658886 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.658946 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:11.158926248 +0000 UTC m=+21.197507418 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.659178 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.659214 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.659230 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:10 crc kubenswrapper[4926]: E1007 20:56:10.659282 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:11.159263248 +0000 UTC m=+21.197844398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.659491 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.659650 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.660574 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.661786 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.661898 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.661801 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.662614 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.662692 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.663118 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.663169 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.663326 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.663523 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.664058 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.664099 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.668806 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.668890 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.669218 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.669427 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.669449 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.670554 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.671106 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.671408 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678062 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678108 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678293 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678396 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678405 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.678443 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.679180 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.679372 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.679463 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.679804 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.679876 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680049 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680120 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680138 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680479 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680484 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680612 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.680621 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.681029 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.681206 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.682932 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.684289 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.684526 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.684676 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.684839 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.685361 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.685802 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.686712 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.686938 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.687059 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.687210 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.687340 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.688288 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.691520 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.692025 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.692373 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.692522 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.692587 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.692645 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.692695 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.693067 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.693619 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.693639 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.693622 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.693974 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.694350 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.694665 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.695004 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.695504 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.696416 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.696878 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.696983 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.697290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.698115 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.698626 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.699713 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.701735 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.702761 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.703532 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.703600 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.707413 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.708419 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.709734 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.710478 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.711091 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.712354 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.713285 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.714259 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.714844 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.714973 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.716926 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.717853 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.718461 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.718988 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.720776 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.721709 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.722511 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.723130 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.723952 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724066 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724090 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724103 4926 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724118 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724129 4926 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724140 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724151 4926 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724161 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724172 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724183 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724211 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724223 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724235 4926 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724247 4926 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724258 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724269 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724279 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724290 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724301 4926 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724311 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724323 4926 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724337 4926 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724351 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724364 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724379 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724392 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724407 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724421 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724436 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724451 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724467 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724485 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724500 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724511 4926 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724523 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724535 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724547 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724559 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724570 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724581 4926 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724854 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724869 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724882 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724897 4926 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724911 4926 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724922 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724934 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724945 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724967 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724979 4926 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.724990 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725004 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725015 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725027 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725038 4926 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725049 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725060 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725070 4926 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725082 4926 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725093 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725106 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725119 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725131 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725141 4926 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725151 4926 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725162 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725173 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725184 4926 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725214 4926 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725225 4926 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725237 4926 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725249 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725260 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725271 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725281 4926 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725292 4926 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725304 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725315 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725326 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725337 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725347 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725359 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725370 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725380 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725391 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725402 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725413 4926 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725426 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725436 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725446 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725457 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725468 4926 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725478 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725488 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725499 4926 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725510 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725522 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725537 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725551 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725566 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725580 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725595 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725608 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725624 4926 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725644 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725658 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725672 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725686 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725701 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725716 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725730 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725746 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725761 4926 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725773 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725784 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725795 4926 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725833 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725844 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725855 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725866 4926 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725912 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725941 4926 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725959 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725976 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.725993 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726013 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726033 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726042 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726051 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726067 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726081 4926 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726096 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726124 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726140 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726158 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726174 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726188 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726229 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726246 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726261 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726279 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726298 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726314 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726329 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726344 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726358 4926 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726373 4926 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726389 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726405 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726734 4926 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726772 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726787 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726800 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726811 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726822 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726833 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726845 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726856 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726867 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726880 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726896 4926 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726910 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726921 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726933 4926 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726944 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726967 4926 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726978 4926 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.726991 4926 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727003 4926 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727017 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727030 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727041 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727052 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727063 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727074 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727085 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727096 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727107 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727119 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727130 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727141 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727163 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727177 4926 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727216 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727229 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727242 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727255 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727267 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727280 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727292 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.727347 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.728881 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.729100 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.729559 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.730031 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.731289 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.731315 4926 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.731484 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.734311 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.735149 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.736144 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.738174 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.738857 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.739868 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.739976 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.740680 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.742324 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.743398 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.744417 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.744997 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.745143 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.745951 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.746527 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.747408 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.747908 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.749020 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.749664 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.750531 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.750965 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.752104 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.752828 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.753366 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.759227 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.770885 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.779862 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.791254 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.800155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.807179 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.808673 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7" exitCode=255 Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.808699 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7"} Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.813689 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.826092 4926 scope.go:117] "RemoveContainer" containerID="72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.826806 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.827071 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.828017 4926 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.828053 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.828062 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.845255 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.860836 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.872949 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.882502 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.892093 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.905235 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.913401 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.940277 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 20:56:10 crc kubenswrapper[4926]: W1007 20:56:10.952856 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cc4f9e1a6dc69bbffc5e825113b93ed281601e1d1f4dbda4bebeed3bbcfce2bd WatchSource:0}: Error finding container cc4f9e1a6dc69bbffc5e825113b93ed281601e1d1f4dbda4bebeed3bbcfce2bd: Status 404 returned error can't find the container with id cc4f9e1a6dc69bbffc5e825113b93ed281601e1d1f4dbda4bebeed3bbcfce2bd Oct 07 20:56:10 crc kubenswrapper[4926]: I1007 20:56:10.959152 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.130965 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.131168 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:56:12.131137323 +0000 UTC m=+22.169718483 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.232235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.232275 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.232297 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.232317 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232422 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232439 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232452 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232476 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232509 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:12.232495785 +0000 UTC m=+22.271076935 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232481 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232573 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:12.232549756 +0000 UTC m=+22.271130996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232637 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:12.232603458 +0000 UTC m=+22.271184608 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232484 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232663 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232674 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.232712 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:12.23268998 +0000 UTC m=+22.271271130 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.678042 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.678121 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.678251 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:11 crc kubenswrapper[4926]: E1007 20:56:11.678386 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.815507 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.818524 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.819672 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.820940 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b58fac4688650bbd70f6c30a2d5c5aad5b16681d0c37141da70a52f9f8afc82f"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.823543 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.823596 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.823609 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cc4f9e1a6dc69bbffc5e825113b93ed281601e1d1f4dbda4bebeed3bbcfce2bd"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.825404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.825454 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"dde68947feefdcd4d0152b4c4e500f3d658a638d81fa965285a26146447c0e0e"} Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.840340 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.864891 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.889444 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.919319 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.931938 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-rmrts"] Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.932226 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.934886 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.935671 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.936747 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.938060 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.950868 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.968186 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:11 crc kubenswrapper[4926]: I1007 20:56:11.993785 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.037464 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfltx\" (UniqueName: \"kubernetes.io/projected/a38bebbe-fe67-479d-9790-9487f62944b8-kube-api-access-rfltx\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.037514 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a38bebbe-fe67-479d-9790-9487f62944b8-host\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.037529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a38bebbe-fe67-479d-9790-9487f62944b8-serviceca\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.043385 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.071339 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.115697 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.138254 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.138337 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfltx\" (UniqueName: \"kubernetes.io/projected/a38bebbe-fe67-479d-9790-9487f62944b8-kube-api-access-rfltx\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.138372 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a38bebbe-fe67-479d-9790-9487f62944b8-host\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.138387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a38bebbe-fe67-479d-9790-9487f62944b8-serviceca\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.139098 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a38bebbe-fe67-479d-9790-9487f62944b8-host\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.139208 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.139160139 +0000 UTC m=+24.177741289 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.139438 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.140443 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a38bebbe-fe67-479d-9790-9487f62944b8-serviceca\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.160894 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.164401 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfltx\" (UniqueName: \"kubernetes.io/projected/a38bebbe-fe67-479d-9790-9487f62944b8-kube-api-access-rfltx\") pod \"node-ca-rmrts\" (UID: \"a38bebbe-fe67-479d-9790-9487f62944b8\") " pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.177457 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.192548 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.211272 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.239310 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.239360 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.239389 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.239412 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239481 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239515 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239536 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239518 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239561 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239571 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.239549652 +0000 UTC m=+24.278130802 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239580 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239582 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239593 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.239583393 +0000 UTC m=+24.278164643 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239598 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239624 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.239609184 +0000 UTC m=+24.278190334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.239649 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.239643185 +0000 UTC m=+24.278224335 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.248890 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rmrts" Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.275026 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda38bebbe_fe67_479d_9790_9487f62944b8.slice/crio-2e1c9e638b52d216b41e5c0003e8316a823527bbf68ce54ed37d4df9f6848cc6 WatchSource:0}: Error finding container 2e1c9e638b52d216b41e5c0003e8316a823527bbf68ce54ed37d4df9f6848cc6: Status 404 returned error can't find the container with id 2e1c9e638b52d216b41e5c0003e8316a823527bbf68ce54ed37d4df9f6848cc6 Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.452867 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-j4btw"] Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.453236 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.455526 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.455545 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.455544 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.470717 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.482120 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.493775 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.506210 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.515787 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.527642 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.538639 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.541889 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm7md\" (UniqueName: \"kubernetes.io/projected/3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d-kube-api-access-gm7md\") pod \"node-resolver-j4btw\" (UID: \"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\") " pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.542149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d-hosts-file\") pod \"node-resolver-j4btw\" (UID: \"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\") " pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.555224 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.567825 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.642670 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d-hosts-file\") pod \"node-resolver-j4btw\" (UID: \"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\") " pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.642710 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm7md\" (UniqueName: \"kubernetes.io/projected/3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d-kube-api-access-gm7md\") pod \"node-resolver-j4btw\" (UID: \"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\") " pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.642804 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d-hosts-file\") pod \"node-resolver-j4btw\" (UID: \"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\") " pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.667114 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm7md\" (UniqueName: \"kubernetes.io/projected/3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d-kube-api-access-gm7md\") pod \"node-resolver-j4btw\" (UID: \"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\") " pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.677950 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.678092 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.682438 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.683208 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.764656 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j4btw" Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.775272 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f15f8f8_8603_4ecd_9035_3aeb61ab3d8d.slice/crio-6fa175719529a8469e0dae9e84204adb7e0ad1c9d88364fb0daa5e76d0a42e7b WatchSource:0}: Error finding container 6fa175719529a8469e0dae9e84204adb7e0ad1c9d88364fb0daa5e76d0a42e7b: Status 404 returned error can't find the container with id 6fa175719529a8469e0dae9e84204adb7e0ad1c9d88364fb0daa5e76d0a42e7b Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.828431 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rmrts" event={"ID":"a38bebbe-fe67-479d-9790-9487f62944b8","Type":"ContainerStarted","Data":"d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f"} Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.828474 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rmrts" event={"ID":"a38bebbe-fe67-479d-9790-9487f62944b8","Type":"ContainerStarted","Data":"2e1c9e638b52d216b41e5c0003e8316a823527bbf68ce54ed37d4df9f6848cc6"} Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.829579 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j4btw" event={"ID":"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d","Type":"ContainerStarted","Data":"6fa175719529a8469e0dae9e84204adb7e0ad1c9d88364fb0daa5e76d0a42e7b"} Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.847527 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.851916 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-t5cd6"] Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.852231 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2zw6c"] Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.852261 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.853135 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2tlv9"] Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.853295 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.853563 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.854052 4926 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.854107 4926 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.854110 4926 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.854060 4926 reflector.go:561] object-"openshift-machine-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.854136 4926 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.854147 4926 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.854165 4926 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.854141 4926 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 20:56:12 crc kubenswrapper[4926]: W1007 20:56:12.854296 4926 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 07 20:56:12 crc kubenswrapper[4926]: E1007 20:56:12.854335 4926 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.855658 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.856226 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.856274 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.856336 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.858055 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.858075 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.858135 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.858410 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.858523 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.859078 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.859125 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.859594 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.862700 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xtr42"] Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.863503 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.864548 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.865253 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.865401 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.878257 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.888569 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.899786 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.907968 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.918122 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.930445 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.942994 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944427 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-daemon-config\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944464 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-socket-dir-parent\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944482 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-config\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-ovn\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944512 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-system-cni-dir\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68x4z\" (UniqueName: \"kubernetes.io/projected/b1217995-75cc-4fc7-8f5f-7752d2e378e2-kube-api-access-68x4z\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944561 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-node-log\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944638 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-bin\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944681 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-k8s-cni-cncf-io\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944738 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-etc-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944761 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-cnibin\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944780 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-kubelet\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944857 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-slash\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944893 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944911 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-log-socket\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944927 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-ovn-kubernetes\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944944 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksxsh\" (UniqueName: \"kubernetes.io/projected/a6eafe65-074f-4285-b66d-ce84d41ee711-kube-api-access-ksxsh\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944966 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-etc-kubernetes\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.944998 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-rootfs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945014 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb2vs\" (UniqueName: \"kubernetes.io/projected/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-kube-api-access-cb2vs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945030 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-conf-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945044 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-netns\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovn-node-metrics-cert\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945081 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-os-release\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945115 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945129 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a6eafe65-074f-4285-b66d-ce84d41ee711-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945150 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-netns\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945183 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-cni-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945214 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b1217995-75cc-4fc7-8f5f-7752d2e378e2-cni-binary-copy\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945234 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-cni-multus\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945248 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-env-overrides\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-script-lib\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-system-cni-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945303 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-cni-bin\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-hostroot\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945340 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-kubelet\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945357 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-var-lib-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945373 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc8t4\" (UniqueName: \"kubernetes.io/projected/9b8c34cd-2651-4f25-8439-1897f0e937b4-kube-api-access-dc8t4\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945388 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a6eafe65-074f-4285-b66d-ce84d41ee711-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945406 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-os-release\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945457 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-multus-certs\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945489 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-netd\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945512 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-proxy-tls\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-systemd-units\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945565 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-cnibin\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-systemd\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.945615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.953497 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.964404 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.982142 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.989818 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.994867 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.996259 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:12Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:12 crc kubenswrapper[4926]: I1007 20:56:12.997798 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.008108 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.018551 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.028301 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.040985 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046382 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-proxy-tls\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046414 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-systemd-units\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046431 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-cnibin\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-systemd\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046476 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-daemon-config\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046526 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-socket-dir-parent\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046560 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-systemd-units\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046614 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-cnibin\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046627 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046584 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-config\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-systemd\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046680 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-socket-dir-parent\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046874 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-ovn\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046932 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-ovn\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.046963 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-system-cni-dir\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047006 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68x4z\" (UniqueName: \"kubernetes.io/projected/b1217995-75cc-4fc7-8f5f-7752d2e378e2-kube-api-access-68x4z\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047029 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-node-log\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047054 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-bin\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047079 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-k8s-cni-cncf-io\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047092 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-node-log\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047133 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-bin\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047140 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-etc-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047101 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-system-cni-dir\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-cnibin\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047145 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-k8s-cni-cncf-io\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047218 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-kubelet\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047242 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-cnibin\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-slash\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047251 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-kubelet\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-etc-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047274 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-config\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047279 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-log-socket\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047327 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-ovn-kubernetes\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047340 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-log-socket\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047386 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-ovn-kubernetes\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047350 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksxsh\" (UniqueName: \"kubernetes.io/projected/a6eafe65-074f-4285-b66d-ce84d41ee711-kube-api-access-ksxsh\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047276 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-slash\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047435 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-etc-kubernetes\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047443 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-daemon-config\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047463 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-rootfs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047486 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb2vs\" (UniqueName: \"kubernetes.io/projected/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-kube-api-access-cb2vs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-conf-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047532 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-netns\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047551 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-conf-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047554 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovn-node-metrics-cert\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047590 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-os-release\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047609 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a6eafe65-074f-4285-b66d-ce84d41ee711-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047649 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-netns\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-cni-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047706 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b1217995-75cc-4fc7-8f5f-7752d2e378e2-cni-binary-copy\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-netns\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047757 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-cni-multus\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-env-overrides\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047805 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-script-lib\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047828 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-system-cni-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-cni-bin\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047873 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-hostroot\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047899 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-kubelet\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-var-lib-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047943 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc8t4\" (UniqueName: \"kubernetes.io/projected/9b8c34cd-2651-4f25-8439-1897f0e937b4-kube-api-access-dc8t4\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047966 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a6eafe65-074f-4285-b66d-ce84d41ee711-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047987 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-os-release\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048008 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-multus-certs\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-rootfs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047510 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-etc-kubernetes\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048030 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-netd\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048076 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-netd\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048133 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-cni-bin\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048167 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-os-release\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048174 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-kubelet\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048183 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-hostroot\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.047787 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-multus-cni-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048258 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-os-release\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048444 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-var-lib-openvswitch\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048501 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-netns\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-var-lib-cni-multus\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048622 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-env-overrides\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048660 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a6eafe65-074f-4285-b66d-ce84d41ee711-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048671 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-host-run-multus-certs\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048703 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1217995-75cc-4fc7-8f5f-7752d2e378e2-system-cni-dir\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048759 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a6eafe65-074f-4285-b66d-ce84d41ee711-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048770 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b1217995-75cc-4fc7-8f5f-7752d2e378e2-cni-binary-copy\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.048835 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a6eafe65-074f-4285-b66d-ce84d41ee711-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.049085 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-script-lib\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.051794 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovn-node-metrics-cert\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.054483 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.062292 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68x4z\" (UniqueName: \"kubernetes.io/projected/b1217995-75cc-4fc7-8f5f-7752d2e378e2-kube-api-access-68x4z\") pod \"multus-2tlv9\" (UID: \"b1217995-75cc-4fc7-8f5f-7752d2e378e2\") " pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.068883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc8t4\" (UniqueName: \"kubernetes.io/projected/9b8c34cd-2651-4f25-8439-1897f0e937b4-kube-api-access-dc8t4\") pod \"ovnkube-node-2zw6c\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.070598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksxsh\" (UniqueName: \"kubernetes.io/projected/a6eafe65-074f-4285-b66d-ce84d41ee711-kube-api-access-ksxsh\") pod \"multus-additional-cni-plugins-xtr42\" (UID: \"a6eafe65-074f-4285-b66d-ce84d41ee711\") " pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.072894 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.085165 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.097371 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.116082 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.127926 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.152141 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.163391 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.170443 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2tlv9" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.173889 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.176477 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.182615 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xtr42" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.188942 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.201977 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: W1007 20:56:13.204563 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b8c34cd_2651_4f25_8439_1897f0e937b4.slice/crio-af39831457da4008b1b1a0ae8a2443ec8ddd724bbf7dd32621981b402e512d0b WatchSource:0}: Error finding container af39831457da4008b1b1a0ae8a2443ec8ddd724bbf7dd32621981b402e512d0b: Status 404 returned error can't find the container with id af39831457da4008b1b1a0ae8a2443ec8ddd724bbf7dd32621981b402e512d0b Oct 07 20:56:13 crc kubenswrapper[4926]: W1007 20:56:13.207167 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6eafe65_074f_4285_b66d_ce84d41ee711.slice/crio-4473c30f92f27d3d83d8c9f8e750622be9bb5a9152131b088e116e855aa8f4d8 WatchSource:0}: Error finding container 4473c30f92f27d3d83d8c9f8e750622be9bb5a9152131b088e116e855aa8f4d8: Status 404 returned error can't find the container with id 4473c30f92f27d3d83d8c9f8e750622be9bb5a9152131b088e116e855aa8f4d8 Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.218675 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.231940 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.251242 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.264959 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.276165 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.296092 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.316576 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.332463 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.465593 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.479681 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.482843 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.493759 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.507244 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.525814 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.542024 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.552681 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.565371 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.584654 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.602282 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.616476 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.634115 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.652477 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.666021 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.678375 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.678454 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:13 crc kubenswrapper[4926]: E1007 20:56:13.678905 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:13 crc kubenswrapper[4926]: E1007 20:56:13.679101 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.683327 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.703276 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.717900 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.730041 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.741796 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.755148 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.769021 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.796485 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.817317 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.833734 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.835224 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerStarted","Data":"8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.835272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerStarted","Data":"ab386719c0c827a8628005e0fb3a7a5c0b22e59dd5b6d7cb14f1e17a3f66bb1e"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.836382 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j4btw" event={"ID":"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d","Type":"ContainerStarted","Data":"4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.837657 4926 generic.go:334] "Generic (PLEG): container finished" podID="a6eafe65-074f-4285-b66d-ce84d41ee711" containerID="a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd" exitCode=0 Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.837719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerDied","Data":"a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.837738 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerStarted","Data":"4473c30f92f27d3d83d8c9f8e750622be9bb5a9152131b088e116e855aa8f4d8"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.839348 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" exitCode=0 Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.839495 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.839548 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"af39831457da4008b1b1a0ae8a2443ec8ddd724bbf7dd32621981b402e512d0b"} Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.846188 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: E1007 20:56:13.851669 4926 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.863704 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.890052 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.930491 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.970896 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:13Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:13 crc kubenswrapper[4926]: I1007 20:56:13.981906 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.034707 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.046881 4926 secret.go:188] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.046962 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-proxy-tls podName:f78a9d2f-35fd-4558-a134-5c8fe26b0a4a nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.546943237 +0000 UTC m=+24.585524387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-proxy-tls") pod "machine-config-daemon-t5cd6" (UID: "f78a9d2f-35fd-4558-a134-5c8fe26b0a4a") : failed to sync secret cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.047960 4926 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.048037 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-mcd-auth-proxy-config podName:f78a9d2f-35fd-4558-a134-5c8fe26b0a4a nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.548018428 +0000 UTC m=+24.586599578 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-mcd-auth-proxy-config") pod "machine-config-daemon-t5cd6" (UID: "f78a9d2f-35fd-4558-a134-5c8fe26b0a4a") : failed to sync configmap cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.065279 4926 projected.go:288] Couldn't get configMap openshift-machine-config-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.065337 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cb2vs for pod openshift-machine-config-operator/machine-config-daemon-t5cd6: failed to sync configmap cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.065399 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-kube-api-access-cb2vs podName:f78a9d2f-35fd-4558-a134-5c8fe26b0a4a nodeName:}" failed. No retries permitted until 2025-10-07 20:56:14.565380442 +0000 UTC m=+24.603961602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cb2vs" (UniqueName: "kubernetes.io/projected/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-kube-api-access-cb2vs") pod "machine-config-daemon-t5cd6" (UID: "f78a9d2f-35fd-4558-a134-5c8fe26b0a4a") : failed to sync configmap cache: timed out waiting for the condition Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.070870 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.081866 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.139869 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.141345 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.160519 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.160672 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:56:18.160647937 +0000 UTC m=+28.199229087 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.161529 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.208884 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.248397 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.261212 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.261263 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.261281 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.261304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261353 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261436 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:18.261415102 +0000 UTC m=+28.299996252 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261447 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261463 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261474 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261472 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261486 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261513 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:18.261499404 +0000 UTC m=+28.300080554 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261507 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261530 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261550 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:18.261544945 +0000 UTC m=+28.300126095 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.261576 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:18.261554766 +0000 UTC m=+28.300136016 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.291344 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.330090 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.381414 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.384384 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.430261 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.469017 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.512078 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.561539 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.564878 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-proxy-tls\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.564989 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.565854 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.569818 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-proxy-tls\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.595065 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.633215 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.666106 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb2vs\" (UniqueName: \"kubernetes.io/projected/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-kube-api-access-cb2vs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.671103 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb2vs\" (UniqueName: \"kubernetes.io/projected/f78a9d2f-35fd-4558-a134-5c8fe26b0a4a-kube-api-access-cb2vs\") pod \"machine-config-daemon-t5cd6\" (UID: \"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\") " pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.675536 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.678876 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:14 crc kubenswrapper[4926]: E1007 20:56:14.679031 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.727439 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.754812 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.796177 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.845832 4926 generic.go:334] "Generic (PLEG): container finished" podID="a6eafe65-074f-4285-b66d-ce84d41ee711" containerID="dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4" exitCode=0 Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.845891 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerDied","Data":"dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.850637 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.850688 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.850705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.850724 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.850737 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.850749 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.865004 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.880780 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.915849 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.955109 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.964053 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:56:14 crc kubenswrapper[4926]: W1007 20:56:14.975893 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf78a9d2f_35fd_4558_a134_5c8fe26b0a4a.slice/crio-b0d80013a3c1254a648ea80cbe442d2654bbb045127342ce1f79e86c2edbda66 WatchSource:0}: Error finding container b0d80013a3c1254a648ea80cbe442d2654bbb045127342ce1f79e86c2edbda66: Status 404 returned error can't find the container with id b0d80013a3c1254a648ea80cbe442d2654bbb045127342ce1f79e86c2edbda66 Oct 07 20:56:14 crc kubenswrapper[4926]: I1007 20:56:14.995876 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:14Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.028898 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.068396 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.109625 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.150581 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.193606 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.231001 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.274282 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.315166 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.351990 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.392125 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.678275 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.678326 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:15 crc kubenswrapper[4926]: E1007 20:56:15.678419 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:15 crc kubenswrapper[4926]: E1007 20:56:15.678549 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.856834 4926 generic.go:334] "Generic (PLEG): container finished" podID="a6eafe65-074f-4285-b66d-ce84d41ee711" containerID="5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1" exitCode=0 Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.856876 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerDied","Data":"5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1"} Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.859099 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1"} Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.859174 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327"} Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.859244 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"b0d80013a3c1254a648ea80cbe442d2654bbb045127342ce1f79e86c2edbda66"} Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.891308 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.915758 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.929716 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.949388 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.970426 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:15 crc kubenswrapper[4926]: I1007 20:56:15.989125 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:15Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.006547 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.029882 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.046262 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.060503 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.077135 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.089585 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.101480 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.114954 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.134876 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.150438 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.170469 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.189977 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.205513 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.217044 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.227143 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.272012 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.315934 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.359858 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.461508 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.486149 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.502827 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.526533 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.559869 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.595236 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.677768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:16 crc kubenswrapper[4926]: E1007 20:56:16.677906 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.864473 4926 generic.go:334] "Generic (PLEG): container finished" podID="a6eafe65-074f-4285-b66d-ce84d41ee711" containerID="497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6" exitCode=0 Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.864577 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerDied","Data":"497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6"} Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.869947 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.898000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.911757 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.915284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.916457 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.916468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.916572 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.916584 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.925772 4926 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.926129 4926 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.927553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.927623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.927646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.927676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.927756 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:16Z","lastTransitionTime":"2025-10-07T20:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.929256 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.947491 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: E1007 20:56:16.957681 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.961406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.961428 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.961436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.961450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.961459 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:16Z","lastTransitionTime":"2025-10-07T20:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.962415 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: E1007 20:56:16.973453 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.982707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.982739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.982748 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.982763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.982775 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:16Z","lastTransitionTime":"2025-10-07T20:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:16 crc kubenswrapper[4926]: I1007 20:56:16.987947 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:16 crc kubenswrapper[4926]: E1007 20:56:16.999479 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.001627 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:16Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.003228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.003260 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.003271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.003297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.003308 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: E1007 20:56:17.016688 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.021509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.021546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.021555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.021571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.021581 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.022213 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: E1007 20:56:17.039823 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: E1007 20:56:17.040005 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.042067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.042100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.042109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.042124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.042133 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.043303 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.056315 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.070626 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.113863 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.143580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.143610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.143620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.143635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.143646 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.150303 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.189722 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.237998 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.246461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.246509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.246523 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.246544 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.246553 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.352232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.352276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.352288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.352306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.352317 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.454828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.454862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.454872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.454891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.454901 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.557896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.557959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.557977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.558002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.558019 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.661080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.661114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.661123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.661137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.661147 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.678473 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.678572 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:17 crc kubenswrapper[4926]: E1007 20:56:17.678650 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:17 crc kubenswrapper[4926]: E1007 20:56:17.678800 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.764450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.764520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.764538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.764568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.764587 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.867323 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.867378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.867393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.867414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.867429 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.877107 4926 generic.go:334] "Generic (PLEG): container finished" podID="a6eafe65-074f-4285-b66d-ce84d41ee711" containerID="1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009" exitCode=0 Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.877154 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerDied","Data":"1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.902637 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.920491 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.937984 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.956423 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.971189 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.971336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.971358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.971382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.971399 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:17Z","lastTransitionTime":"2025-10-07T20:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.972928 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:17 crc kubenswrapper[4926]: I1007 20:56:17.989579 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:17Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.009654 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.041377 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.063644 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.074652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.074686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.074695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.074710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.074721 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.078228 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.093755 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.107886 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.122452 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.137155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.157097 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.177933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.177965 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.177972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.177986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.177996 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.201582 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.201952 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:56:26.201872964 +0000 UTC m=+36.240454144 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.281069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.281130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.281147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.281169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.281186 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.302779 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.302870 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.302939 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.303012 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303019 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303072 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303121 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303153 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303152 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303165 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303247 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303268 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303153 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:26.303129763 +0000 UTC m=+36.341710943 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303328 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:26.303295697 +0000 UTC m=+36.341876967 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303363 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:26.303345359 +0000 UTC m=+36.341926699 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.303409 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:26.30339192 +0000 UTC m=+36.341973200 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.384495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.384550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.384567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.384590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.384606 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.488896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.488941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.488954 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.488971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.488982 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.592778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.592835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.592854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.592872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.592884 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.685359 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:18 crc kubenswrapper[4926]: E1007 20:56:18.685523 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.698951 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.699007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.699025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.699049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.699065 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.801513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.802528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.802555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.802580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.802597 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.884400 4926 generic.go:334] "Generic (PLEG): container finished" podID="a6eafe65-074f-4285-b66d-ce84d41ee711" containerID="11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552" exitCode=0 Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.884477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerDied","Data":"11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.904766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.904792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.904801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.904815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.904825 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:18Z","lastTransitionTime":"2025-10-07T20:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.905868 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.941531 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.955732 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.969940 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:18 crc kubenswrapper[4926]: I1007 20:56:18.988776 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:18Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.005535 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.008573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.008606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.008618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.008637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.008649 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.021667 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.036498 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.053307 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.083550 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.105296 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.111215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.111254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.111262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.111297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.111309 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.119033 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.138519 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.157704 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.178313 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.214210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.214240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.214248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.214262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.214271 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.317155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.317221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.317235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.317253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.317265 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.420026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.420083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.420101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.420125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.420141 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.523261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.523319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.523336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.523358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.523375 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.626370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.626434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.626451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.626476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.626497 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.678636 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.678759 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:19 crc kubenswrapper[4926]: E1007 20:56:19.678821 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:19 crc kubenswrapper[4926]: E1007 20:56:19.678992 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.729489 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.729548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.729567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.729594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.729613 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.833233 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.833298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.833317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.833343 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.833363 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.895508 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" event={"ID":"a6eafe65-074f-4285-b66d-ce84d41ee711","Type":"ContainerStarted","Data":"2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.903064 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.903459 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.903499 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.938757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.938816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.938834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.938858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.938877 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:19Z","lastTransitionTime":"2025-10-07T20:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.955274 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.963247 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.965081 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.988362 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:19 crc kubenswrapper[4926]: I1007 20:56:19.999441 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.014065 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.024951 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.034892 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.041399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.041436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.041446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.041462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.041475 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.046953 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.076157 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.085246 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.103155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.115647 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.130307 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.142802 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.143771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.143806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.143816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.143835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.143847 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.154369 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.167325 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.182595 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.195107 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.211027 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.225030 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.238607 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.246314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.246358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.246370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.246388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.246399 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.254000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.273384 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.302942 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.318172 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.332266 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.346124 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.348717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.348763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.348782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.348802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.348816 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.360625 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.376261 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.389836 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.408899 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.452158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.452256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.452281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.452312 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.452333 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.555621 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.555684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.555705 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.555731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.555749 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.659180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.659273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.659290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.659316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.659335 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.678903 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:20 crc kubenswrapper[4926]: E1007 20:56:20.679122 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.701757 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.732946 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.749697 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.761507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.761584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.761603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.761627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.761645 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.771406 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.793973 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.821345 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.841251 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.858274 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.863894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.863962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.863982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.864008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.864029 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.877046 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.907519 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.909753 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.928937 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.946426 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.967163 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.967250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.967272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.967300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.967325 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:20Z","lastTransitionTime":"2025-10-07T20:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.968607 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:20 crc kubenswrapper[4926]: I1007 20:56:20.991118 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.007711 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:21Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.070828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.070892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.070916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.070957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.070979 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.174228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.174286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.174304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.174330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.174351 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.277493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.277546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.277561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.277583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.277597 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.379907 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.379964 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.379977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.379992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.380004 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.482626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.482661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.482672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.482690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.482701 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.584796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.584825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.584833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.584847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.584856 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.678576 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.678606 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:21 crc kubenswrapper[4926]: E1007 20:56:21.679047 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:21 crc kubenswrapper[4926]: E1007 20:56:21.678790 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.687634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.687687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.687706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.687728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.687745 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.792325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.792671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.792688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.792713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.792730 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.895111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.895166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.895179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.895220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.895237 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.916044 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.997172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.997240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.997254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.997271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:21 crc kubenswrapper[4926]: I1007 20:56:21.997282 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:21Z","lastTransitionTime":"2025-10-07T20:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.100206 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.100244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.100254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.100267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.100276 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.203792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.203852 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.203872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.203896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.203917 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.307739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.307797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.307815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.307838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.307856 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.410549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.410600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.410616 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.410640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.410658 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.513553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.513624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.513638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.513657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.513670 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.615903 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.615945 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.615955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.615972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.616166 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.677828 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:22 crc kubenswrapper[4926]: E1007 20:56:22.678018 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.718966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.719054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.719080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.719113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.719139 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.822186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.822271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.822295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.822327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.822348 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.922174 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/0.log" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.924566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.924620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.924637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.924660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.924677 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:22Z","lastTransitionTime":"2025-10-07T20:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.926532 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3" exitCode=1 Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.926595 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3"} Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.927894 4926 scope.go:117] "RemoveContainer" containerID="c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.951528 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:22Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.976110 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:22Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:22 crc kubenswrapper[4926]: I1007 20:56:22.996823 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:22Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.019515 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.027415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.027464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.027484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.027506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.027524 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.050731 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:21Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 20:56:21.940709 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 20:56:21.940795 6228 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:21.940809 6228 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 20:56:21.940852 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 20:56:21.940850 6228 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:21.940877 6228 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 20:56:21.940886 6228 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 20:56:21.940912 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 20:56:21.940920 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 20:56:21.940948 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 20:56:21.940966 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 20:56:21.940987 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 20:56:21.940955 6228 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.077540 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.099903 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.123370 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.130037 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.130097 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.130119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.130147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.130167 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.145464 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.159806 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.176481 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.199449 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.227930 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.235001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.235062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.235082 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.235110 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.235133 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.248038 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.267873 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.337587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.337649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.337669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.337692 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.337709 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.400574 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.440663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.440715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.440733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.440758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.440777 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.544319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.544386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.544411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.544440 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.544463 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.647504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.647596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.647620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.647651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.647673 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.678239 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.678248 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:23 crc kubenswrapper[4926]: E1007 20:56:23.678532 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:23 crc kubenswrapper[4926]: E1007 20:56:23.678664 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.750980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.751039 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.751057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.751082 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.751106 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.854047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.854092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.854106 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.854125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.854137 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.934842 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/0.log" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.938816 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.939409 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.956769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.956824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.956844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.956869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.956886 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:23Z","lastTransitionTime":"2025-10-07T20:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.959938 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:23 crc kubenswrapper[4926]: I1007 20:56:23.983411 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:21Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 20:56:21.940709 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 20:56:21.940795 6228 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:21.940809 6228 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 20:56:21.940852 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 20:56:21.940850 6228 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:21.940877 6228 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 20:56:21.940886 6228 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 20:56:21.940912 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 20:56:21.940920 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 20:56:21.940948 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 20:56:21.940966 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 20:56:21.940987 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 20:56:21.940955 6228 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:23Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.003170 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.020568 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.038666 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.054435 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.060158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.060268 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.060293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.060326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.060361 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.072815 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.091924 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.110271 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.127379 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.161363 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.163293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.163369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.163392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.163420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.163439 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.183592 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.206341 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.231881 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.253375 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.266931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.266993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.267017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.267047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.267068 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.370084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.370139 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.370156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.370185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.370241 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.472690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.472749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.472769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.472793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.472811 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.585314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.585591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.585681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.585772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.585851 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.678140 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:24 crc kubenswrapper[4926]: E1007 20:56:24.678396 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.688646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.688707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.688753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.688778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.688795 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.792401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.792472 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.792492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.792516 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.792539 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.895167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.895257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.895277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.895301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.895319 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.945504 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/1.log" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.946358 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/0.log" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.950882 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e" exitCode=1 Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.950962 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e"} Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.951051 4926 scope.go:117] "RemoveContainer" containerID="c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.951882 4926 scope.go:117] "RemoveContainer" containerID="cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e" Oct 07 20:56:24 crc kubenswrapper[4926]: E1007 20:56:24.952136 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.987562 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:24Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.998243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.998325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.998364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.998397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:24 crc kubenswrapper[4926]: I1007 20:56:24.998421 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:24Z","lastTransitionTime":"2025-10-07T20:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.011291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.028801 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.051900 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.071939 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.092245 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.101533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.101593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.101620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.101648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.101670 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.112749 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.142913 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:21Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 20:56:21.940709 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 20:56:21.940795 6228 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:21.940809 6228 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 20:56:21.940852 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 20:56:21.940850 6228 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:21.940877 6228 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 20:56:21.940886 6228 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 20:56:21.940912 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 20:56:21.940920 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 20:56:21.940948 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 20:56:21.940966 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 20:56:21.940987 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 20:56:21.940955 6228 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.159954 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.180612 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.204704 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.205387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.205441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.205466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.205497 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.205519 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.225787 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.245068 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.264683 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.283896 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.309024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.309067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.309085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.309110 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.309127 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.399341 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl"] Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.400170 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.404543 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.405837 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.411905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.411959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.411979 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.412002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.412021 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.419500 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.442412 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.465848 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.485458 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7cea4c9a-b58a-4d94-91ad-4056b318811a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.485536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7cea4c9a-b58a-4d94-91ad-4056b318811a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.485580 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzddl\" (UniqueName: \"kubernetes.io/projected/7cea4c9a-b58a-4d94-91ad-4056b318811a-kube-api-access-qzddl\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.485641 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7cea4c9a-b58a-4d94-91ad-4056b318811a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.489831 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.509400 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.514759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.514820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.514839 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.514863 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.514879 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.527152 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.544016 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.578797 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.586837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7cea4c9a-b58a-4d94-91ad-4056b318811a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.586968 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7cea4c9a-b58a-4d94-91ad-4056b318811a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.587007 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7cea4c9a-b58a-4d94-91ad-4056b318811a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.587046 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzddl\" (UniqueName: \"kubernetes.io/projected/7cea4c9a-b58a-4d94-91ad-4056b318811a-kube-api-access-qzddl\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.588287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7cea4c9a-b58a-4d94-91ad-4056b318811a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.588474 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7cea4c9a-b58a-4d94-91ad-4056b318811a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.595914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7cea4c9a-b58a-4d94-91ad-4056b318811a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.599534 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.616699 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzddl\" (UniqueName: \"kubernetes.io/projected/7cea4c9a-b58a-4d94-91ad-4056b318811a-kube-api-access-qzddl\") pod \"ovnkube-control-plane-749d76644c-v6dbl\" (UID: \"7cea4c9a-b58a-4d94-91ad-4056b318811a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.617376 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.617872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.617925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.617939 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.617962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.617975 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.635857 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.657782 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.673339 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.677802 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.677867 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:25 crc kubenswrapper[4926]: E1007 20:56:25.677952 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:25 crc kubenswrapper[4926]: E1007 20:56:25.678049 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.692755 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.713424 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.720109 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.720638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.720693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.720718 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.720747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.720768 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: W1007 20:56:25.741799 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cea4c9a_b58a_4d94_91ad_4056b318811a.slice/crio-e8d2502a7df0224b83e20b19c0a6098694dd69d4b7b85876506811724887c0c0 WatchSource:0}: Error finding container e8d2502a7df0224b83e20b19c0a6098694dd69d4b7b85876506811724887c0c0: Status 404 returned error can't find the container with id e8d2502a7df0224b83e20b19c0a6098694dd69d4b7b85876506811724887c0c0 Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.753809 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4aa6b54a80aef489e2475caf60d5447bef0bcd1e2e5cd5132e312945aa2a2b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:21Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 20:56:21.940709 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 20:56:21.940795 6228 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:21.940809 6228 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 20:56:21.940831 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 20:56:21.940852 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 20:56:21.940850 6228 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:21.940877 6228 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 20:56:21.940886 6228 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 20:56:21.940912 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 20:56:21.940920 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 20:56:21.940948 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 20:56:21.940966 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 20:56:21.940987 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 20:56:21.940955 6228 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.823404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.823440 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.823448 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.823461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.823469 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.925921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.925961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.925970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.925984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.925994 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:25Z","lastTransitionTime":"2025-10-07T20:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.958744 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/1.log" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.962604 4926 scope.go:117] "RemoveContainer" containerID="cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e" Oct 07 20:56:25 crc kubenswrapper[4926]: E1007 20:56:25.962729 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.963696 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" event={"ID":"7cea4c9a-b58a-4d94-91ad-4056b318811a","Type":"ContainerStarted","Data":"a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.963738 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" event={"ID":"7cea4c9a-b58a-4d94-91ad-4056b318811a","Type":"ContainerStarted","Data":"e8d2502a7df0224b83e20b19c0a6098694dd69d4b7b85876506811724887c0c0"} Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.983128 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:25 crc kubenswrapper[4926]: I1007 20:56:25.996506 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:25Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.012155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.030053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.030268 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.030503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.030646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.030827 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.032485 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.044980 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.055808 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.066845 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.080602 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.096709 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.119411 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.134212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.134268 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.134283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.134303 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.134323 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.136599 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.160058 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.181521 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.194531 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.206337 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.219655 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.236487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.236531 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.236543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.236560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.236572 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.293589 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.293856 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:56:42.293817143 +0000 UTC m=+52.332398333 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.339921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.339974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.339992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.340016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.340034 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.395260 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.395331 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.395372 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.395422 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395520 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395566 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395615 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395637 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395672 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395703 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395756 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395777 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395591 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:42.395567606 +0000 UTC m=+52.434148766 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395881 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:42.395857065 +0000 UTC m=+52.434438255 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395904 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:42.395892606 +0000 UTC m=+52.434473796 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.395927 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:42.395917536 +0000 UTC m=+52.434498716 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.442807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.442862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.442879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.442904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.442921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.532554 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-c6mr4"] Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.533329 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.533417 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.546083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.546129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.546147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.546170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.546187 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.547918 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.564552 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.585122 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.597488 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.597635 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j5g8\" (UniqueName: \"kubernetes.io/projected/7a8e664c-9b14-4510-8a51-bec46ecbaea5-kube-api-access-5j5g8\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.609168 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.627977 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.643610 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.648507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.648546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.648553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.648565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.648573 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.661335 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.676306 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.678559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.678659 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.698392 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.698747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j5g8\" (UniqueName: \"kubernetes.io/projected/7a8e664c-9b14-4510-8a51-bec46ecbaea5-kube-api-access-5j5g8\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.698921 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.699064 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: E1007 20:56:26.699136 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:27.199120306 +0000 UTC m=+37.237701456 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.713466 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.726995 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j5g8\" (UniqueName: \"kubernetes.io/projected/7a8e664c-9b14-4510-8a51-bec46ecbaea5-kube-api-access-5j5g8\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.728896 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.745929 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.750695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.750731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.750742 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.750759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.750776 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.767122 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.785941 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.803800 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.821645 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.851568 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.852993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.853025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.853036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.853054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.853066 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.957304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.957377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.957400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.957427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.957455 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:26Z","lastTransitionTime":"2025-10-07T20:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.970533 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" event={"ID":"7cea4c9a-b58a-4d94-91ad-4056b318811a","Type":"ContainerStarted","Data":"6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b"} Oct 07 20:56:26 crc kubenswrapper[4926]: I1007 20:56:26.995995 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:26Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.014113 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.029975 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.047738 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.060956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.061000 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.061012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.061029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.061039 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.061100 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.072435 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.086261 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.097998 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.121415 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.139337 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.153852 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.163986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.164035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.164047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.164068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.164082 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.165278 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.177800 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.197972 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.204160 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.204328 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.204434 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:28.204407841 +0000 UTC m=+38.242989001 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.217654 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.232956 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.255764 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.266603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.266654 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.266673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.266696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.266713 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.309959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.310031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.310056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.310091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.310115 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.326315 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.331843 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.331892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.331908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.331931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.331949 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.351092 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.355317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.355369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.355383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.355402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.355415 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.370679 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.374408 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.374449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.374461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.374479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.374492 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.391623 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.396514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.396554 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.396566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.396582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.396594 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.414110 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:27Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.414277 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.416311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.416359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.416371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.416389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.416401 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.519446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.519513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.519538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.519571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.519598 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.623295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.623368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.623391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.623418 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.623436 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.677915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.677992 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.677954 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.678181 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.678333 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:27 crc kubenswrapper[4926]: E1007 20:56:27.678492 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.726580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.726657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.726679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.726709 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.726730 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.830833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.831222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.831375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.831550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.831718 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.935252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.935312 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.935329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.935354 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:27 crc kubenswrapper[4926]: I1007 20:56:27.935373 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:27Z","lastTransitionTime":"2025-10-07T20:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.038178 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.038281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.038308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.038337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.038356 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.141894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.141945 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.141961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.141984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.142001 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.216093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:28 crc kubenswrapper[4926]: E1007 20:56:28.216319 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:28 crc kubenswrapper[4926]: E1007 20:56:28.216415 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:30.216382131 +0000 UTC m=+40.254963321 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.245085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.245156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.245175 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.245236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.245257 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.348854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.348911 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.348937 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.348969 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.348989 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.452655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.453022 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.453167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.453420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.453564 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.556309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.556347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.556357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.556372 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.556381 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.659070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.659140 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.659158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.659261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.659289 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.678708 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:28 crc kubenswrapper[4926]: E1007 20:56:28.678892 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.762290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.762364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.762388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.762414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.762431 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.865491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.865550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.865570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.865593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.865612 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.968468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.968538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.968556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.968580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:28 crc kubenswrapper[4926]: I1007 20:56:28.968598 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:28Z","lastTransitionTime":"2025-10-07T20:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.059703 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.071674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.071717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.071734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.071756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.071775 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.081775 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.105052 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.124913 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.146044 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.164115 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.174464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.174521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.174540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.174571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.174593 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.181178 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.197543 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.214820 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.253177 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.274373 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.278809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.278871 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.278890 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.278918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.278937 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.293445 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.310817 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.332433 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.351528 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.372367 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.381563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.381633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.381657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.381682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.381702 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.395921 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.426867 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.484771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.484840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.484860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.484886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.484904 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.587966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.588036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.588054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.588083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.588101 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.678241 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.678246 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.678349 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:29 crc kubenswrapper[4926]: E1007 20:56:29.678529 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:29 crc kubenswrapper[4926]: E1007 20:56:29.678693 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:29 crc kubenswrapper[4926]: E1007 20:56:29.678841 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.690976 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.691034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.691051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.691074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.691092 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.794442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.794554 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.794579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.794611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.794653 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.896618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.896650 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.896659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.896688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:29 crc kubenswrapper[4926]: I1007 20:56:29.896698 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:29Z","lastTransitionTime":"2025-10-07T20:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.000173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.000292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.000311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.000337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.000354 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.103411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.103507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.103530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.103561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.103587 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.206360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.206424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.206442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.206467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.206484 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.246030 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:30 crc kubenswrapper[4926]: E1007 20:56:30.246272 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:30 crc kubenswrapper[4926]: E1007 20:56:30.246359 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:34.246334675 +0000 UTC m=+44.284915865 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.308514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.308590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.308612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.308642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.308664 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.411451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.411521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.411546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.411578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.411602 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.514671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.514711 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.514720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.514736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.514745 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.617273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.617337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.617363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.617389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.617405 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.678827 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:30 crc kubenswrapper[4926]: E1007 20:56:30.678977 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.701462 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.720451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.720496 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.720508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.720524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.720534 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.727047 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.744815 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.764303 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.790978 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.808132 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.822482 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.822525 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.822537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.822554 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.822566 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.823949 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.844607 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.872279 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.891897 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.909506 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.923254 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.924927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.925012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.925034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.925057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.925074 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:30Z","lastTransitionTime":"2025-10-07T20:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.937783 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.953785 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.969646 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.984254 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:30 crc kubenswrapper[4926]: I1007 20:56:30.999775 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:30Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.028668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.028737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.028755 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.028782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.028800 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.132326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.132379 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.132399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.132423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.132441 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.235578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.235689 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.235736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.235768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.235790 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.338707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.338760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.338777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.338798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.338817 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.442285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.442351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.442370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.442403 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.442422 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.545391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.545469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.545491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.545521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.545543 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.648073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.648138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.648160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.648232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.648271 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.678693 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.678734 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.678739 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:31 crc kubenswrapper[4926]: E1007 20:56:31.679380 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:31 crc kubenswrapper[4926]: E1007 20:56:31.679583 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:31 crc kubenswrapper[4926]: E1007 20:56:31.679185 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.751228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.751300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.751336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.751367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.751388 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.854331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.854391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.854407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.854432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.854450 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.957157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.957248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.957266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.957290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:31 crc kubenswrapper[4926]: I1007 20:56:31.957307 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:31Z","lastTransitionTime":"2025-10-07T20:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.060700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.060756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.060773 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.060798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.060815 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.164025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.164065 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.164074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.164089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.164100 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.267367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.267500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.267560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.267587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.267641 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.371294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.371339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.371351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.371368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.371379 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.474508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.474620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.474677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.474700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.474716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.577964 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.578030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.578046 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.578071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.578093 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.678508 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:32 crc kubenswrapper[4926]: E1007 20:56:32.678934 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.680723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.680786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.680803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.680825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.680843 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.783673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.783732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.783749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.783775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.783791 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.886992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.887053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.887075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.887107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.887127 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.989787 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.989850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.989872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.989901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:32 crc kubenswrapper[4926]: I1007 20:56:32.989921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:32Z","lastTransitionTime":"2025-10-07T20:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.092476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.092542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.092573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.092598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.092616 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.195146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.195287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.195307 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.195331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.195353 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.297533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.297589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.297609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.297634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.297653 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.400434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.400494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.400519 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.400552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.400574 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.503104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.503168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.503229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.503261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.503282 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.605991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.606038 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.606054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.606076 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.606095 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.678004 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.678025 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.678286 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:33 crc kubenswrapper[4926]: E1007 20:56:33.678180 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:33 crc kubenswrapper[4926]: E1007 20:56:33.678433 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:33 crc kubenswrapper[4926]: E1007 20:56:33.678547 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.711154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.711329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.711357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.711389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.711422 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.813818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.813885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.813902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.813925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.813941 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.916822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.917067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.917150 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.917178 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:33 crc kubenswrapper[4926]: I1007 20:56:33.917248 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:33Z","lastTransitionTime":"2025-10-07T20:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.020024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.020160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.020186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.020242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.020265 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.123313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.123367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.123384 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.123411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.123430 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.227383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.227460 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.227485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.227516 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.227539 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.286026 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:34 crc kubenswrapper[4926]: E1007 20:56:34.286334 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:34 crc kubenswrapper[4926]: E1007 20:56:34.286461 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:42.286396988 +0000 UTC m=+52.324978178 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.330838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.330904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.330922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.330947 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.330964 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.433342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.433416 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.433440 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.433467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.433482 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.537080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.537169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.537227 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.537259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.537283 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.640259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.640325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.640352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.640383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.640407 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.678961 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:34 crc kubenswrapper[4926]: E1007 20:56:34.679118 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.743267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.743358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.743377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.743399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.743417 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.846358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.846469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.846487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.846510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.846526 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.949722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.949781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.949799 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.949824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:34 crc kubenswrapper[4926]: I1007 20:56:34.949844 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:34Z","lastTransitionTime":"2025-10-07T20:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.052619 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.052683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.052702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.052730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.052748 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.156429 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.156494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.156511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.156538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.156558 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.259874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.260019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.260051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.260116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.260137 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.363980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.364053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.364073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.364104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.364130 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.466425 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.466495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.466514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.466540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.466558 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.570506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.570567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.570586 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.570611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.570628 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.673906 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.673960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.673983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.674015 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.674037 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.678624 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:35 crc kubenswrapper[4926]: E1007 20:56:35.678774 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.679022 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:35 crc kubenswrapper[4926]: E1007 20:56:35.679127 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.679323 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:35 crc kubenswrapper[4926]: E1007 20:56:35.679660 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.776317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.776363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.776375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.776392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.776406 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.878548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.878601 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.878617 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.878634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.878649 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.981824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.981879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.981896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.981921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:35 crc kubenswrapper[4926]: I1007 20:56:35.981939 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:35Z","lastTransitionTime":"2025-10-07T20:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.085344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.085483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.085508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.085535 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.085557 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.188866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.189134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.189292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.189453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.189678 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.292869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.292930 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.292949 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.292973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.292990 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.397043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.397144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.397173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.397259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.397286 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.500920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.501002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.501022 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.501056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.501077 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.603908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.603992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.604012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.604036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.604053 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.678470 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:36 crc kubenswrapper[4926]: E1007 20:56:36.678651 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.679931 4926 scope.go:117] "RemoveContainer" containerID="cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.707441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.707690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.707972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.708150 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.708379 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.813048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.813103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.813124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.813150 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.813171 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.917006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.917072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.917093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.917122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:36 crc kubenswrapper[4926]: I1007 20:56:36.917142 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:36Z","lastTransitionTime":"2025-10-07T20:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.010703 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/1.log" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.015444 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.016093 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.019384 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.019435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.019454 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.019479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.019497 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.031105 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.052603 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.073172 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.108139 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.121907 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.121968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.121985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.122010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.122027 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.138625 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.166870 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.190446 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.214652 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.224615 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.224664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.224676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.224697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.224709 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.233738 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.246774 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.276486 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.291650 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.302479 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.313968 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.326874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.326914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.326925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.326943 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.326955 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.328829 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.342855 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.356326 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.428618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.428664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.428678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.428694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.428705 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.531570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.531609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.531620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.531637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.531650 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.634332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.634388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.634409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.634433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.634452 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.678757 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.678869 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.678944 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.679079 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.679310 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.679642 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.680127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.680353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.680499 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.680662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.680801 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.699387 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.705005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.705049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.705067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.705088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.705106 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.722235 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.726964 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.727180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.727385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.727529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.727676 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.747865 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.751990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.752022 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.752032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.752048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.752059 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.770499 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.775251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.775283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.775291 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.775305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.775316 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.792943 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:37Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:37 crc kubenswrapper[4926]: E1007 20:56:37.793544 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.795810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.795890 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.795910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.795933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.795952 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.899576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.899638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.899655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.899680 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:37 crc kubenswrapper[4926]: I1007 20:56:37.899701 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:37Z","lastTransitionTime":"2025-10-07T20:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.002934 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.002983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.002996 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.003015 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.003028 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.022357 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/2.log" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.023626 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/1.log" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.027996 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d" exitCode=1 Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.028048 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.028086 4926 scope.go:117] "RemoveContainer" containerID="cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.029641 4926 scope.go:117] "RemoveContainer" containerID="7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d" Oct 07 20:56:38 crc kubenswrapper[4926]: E1007 20:56:38.030135 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.053263 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.087184 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb2468e8c5b13fecd59b12d1f5aa9190d544ce94ae384cfa6680d6c3891c455e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:24Z\\\",\\\"message\\\":\\\" in node crc\\\\nI1007 20:56:24.276368 6372 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1007 20:56:24.276378 6372 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 20:56:24.276289 6372 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1007 20:56:24.276395 6372 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1007 20:56:24.276394 6372 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nF1007 20:56:24.276421 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.106280 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.106369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.106389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.106415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.106433 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.109422 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.131452 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.150336 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.167248 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.185491 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.206781 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.210791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.210985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.211018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.211103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.211159 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.232121 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.251008 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.283336 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.303710 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.314122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.314190 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.314261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.314292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.314312 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.323495 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.340581 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.363319 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.384428 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.405801 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:38Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.417444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.417495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.417514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.417542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.417560 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.520404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.520443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.520453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.520469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.520481 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.623637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.623733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.623750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.623775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.623798 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.678904 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:38 crc kubenswrapper[4926]: E1007 20:56:38.679078 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.726392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.726433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.726443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.726458 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.726470 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.829859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.829932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.829959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.829992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.830016 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.933710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.933764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.933782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.933806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:38 crc kubenswrapper[4926]: I1007 20:56:38.933824 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:38Z","lastTransitionTime":"2025-10-07T20:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.036374 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/2.log" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.037169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.037415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.037443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.037517 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.037623 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.041918 4926 scope.go:117] "RemoveContainer" containerID="7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d" Oct 07 20:56:39 crc kubenswrapper[4926]: E1007 20:56:39.042169 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.064643 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.083937 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.102762 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.139453 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.140706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.140761 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.140779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.140804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.141435 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.160809 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.181101 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.202274 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.232502 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.245307 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.245365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.245383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.245406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.245424 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.259150 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.278951 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.299102 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.316865 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.333889 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.349264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.349316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.349335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.349360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.349379 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.350619 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.372769 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.396486 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.412696 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:39Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.452295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.452359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.452375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.452399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.452418 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.555886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.555948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.555966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.555995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.556014 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.658927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.658992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.659010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.659034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.659051 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.678267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.678287 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:39 crc kubenswrapper[4926]: E1007 20:56:39.678422 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.678510 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:39 crc kubenswrapper[4926]: E1007 20:56:39.678677 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:39 crc kubenswrapper[4926]: E1007 20:56:39.678756 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.762383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.762441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.762457 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.762479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.762494 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.865575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.865635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.865652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.865676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.865692 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.968749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.968798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.968811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.968829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:39 crc kubenswrapper[4926]: I1007 20:56:39.968843 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:39Z","lastTransitionTime":"2025-10-07T20:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.072188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.072283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.072301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.072323 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.072342 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.175099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.175182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.175234 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.175264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.175288 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.278261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.278334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.278352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.278377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.278396 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.381861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.381924 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.381941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.381964 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.381985 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.484882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.484939 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.484956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.484978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.484995 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.588029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.588093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.588111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.588137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.588155 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.678584 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:40 crc kubenswrapper[4926]: E1007 20:56:40.678749 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.689974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.690039 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.690065 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.690099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.690123 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.707033 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.730921 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.751467 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.771092 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.793271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.793357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.793383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.793419 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.793441 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.797012 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.819975 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.840290 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.861673 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.895865 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.897101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.897160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.897177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.897235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.897253 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.912010 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.931716 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.956656 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.978716 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.999683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:40 crc kubenswrapper[4926]: I1007 20:56:40.999747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:40.999770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:40.999797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:40.999881 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:40Z","lastTransitionTime":"2025-10-07T20:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.001422 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:40Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.021291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:41Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.037578 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:41Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.056605 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:41Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.102504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.102559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.102576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.102599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.102618 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.206266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.206321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.206341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.206363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.206380 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.309367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.309426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.309444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.309469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.309490 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.412847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.412911 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.412931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.412956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.412973 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.516344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.516439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.516465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.517002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.517312 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.621609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.621648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.621659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.621676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.621688 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.678744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.678744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:41 crc kubenswrapper[4926]: E1007 20:56:41.678919 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:41 crc kubenswrapper[4926]: E1007 20:56:41.678990 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.678769 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:41 crc kubenswrapper[4926]: E1007 20:56:41.679100 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.723979 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.724031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.724047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.724068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.724084 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.827115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.827172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.827221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.827248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.827265 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.929938 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.929995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.930013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.930037 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:41 crc kubenswrapper[4926]: I1007 20:56:41.930054 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:41Z","lastTransitionTime":"2025-10-07T20:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.032785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.032843 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.032860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.032883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.032900 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.135448 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.135524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.135550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.135580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.135602 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.238293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.238366 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.238388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.238415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.238432 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.340736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.340779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.340790 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.340810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.340821 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.378399 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.378514 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.378572 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:57:14.378540713 +0000 UTC m=+84.417121903 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.378613 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.378664 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:56:58.378649396 +0000 UTC m=+68.417230536 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.444255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.444315 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.444333 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.444357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.444378 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.479617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.479742 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.479793 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.479844 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.479955 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:57:14.479923435 +0000 UTC m=+84.518504625 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.479983 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480014 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.479857 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480057 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480058 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480030 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480108 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480084 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480158 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:57:14.480131151 +0000 UTC m=+84.518712341 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480412 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:57:14.480388759 +0000 UTC m=+84.518969939 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.480455 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:57:14.48043515 +0000 UTC m=+84.519016330 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.547545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.547619 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.547638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.547662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.547682 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.650615 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.650663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.650681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.650704 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.650720 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.678493 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:42 crc kubenswrapper[4926]: E1007 20:56:42.678745 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.753866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.753934 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.753955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.753986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.754004 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.857305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.857338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.857347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.857360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.857401 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.961013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.961092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.961115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.961146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:42 crc kubenswrapper[4926]: I1007 20:56:42.961165 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:42Z","lastTransitionTime":"2025-10-07T20:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.064225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.064322 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.064341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.064369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.064387 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.111934 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.127892 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.134614 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.167854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.167912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.167931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.167955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.167973 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.170565 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.186814 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.206626 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.230591 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.252715 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.267822 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.270860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.270903 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.270920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.270945 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.270965 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.283984 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.300862 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.318604 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.349953 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.368853 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.373024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.373086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.373109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.373138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.373162 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.387679 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.407480 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.434030 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.453933 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.474619 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:43Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.476330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.476390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.476410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.476436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.476454 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.578276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.578326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.578338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.578350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.578359 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.678966 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:43 crc kubenswrapper[4926]: E1007 20:56:43.679126 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.679535 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:43 crc kubenswrapper[4926]: E1007 20:56:43.679855 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.679892 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:43 crc kubenswrapper[4926]: E1007 20:56:43.680126 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.680641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.680666 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.680674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.680691 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.680700 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.783511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.783552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.783563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.783579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.783591 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.886620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.886656 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.886667 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.886683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.886693 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.988601 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.988637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.988647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.988661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:43 crc kubenswrapper[4926]: I1007 20:56:43.988671 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:43Z","lastTransitionTime":"2025-10-07T20:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.091966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.092066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.092091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.092123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.092144 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.195100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.195157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.195174 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.195219 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.195238 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.298009 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.298071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.298087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.298113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.298167 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.400693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.400745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.400763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.400786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.400803 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.502891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.502951 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.502962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.502976 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.503005 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.605071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.605110 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.605119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.605133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.605142 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.678383 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:44 crc kubenswrapper[4926]: E1007 20:56:44.678527 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.707953 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.708036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.708054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.708079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.708096 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.810105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.810187 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.810247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.810271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.810287 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.913159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.913251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.913269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.913295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:44 crc kubenswrapper[4926]: I1007 20:56:44.913311 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:44Z","lastTransitionTime":"2025-10-07T20:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.016019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.016083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.016108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.016138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.016162 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.118814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.118872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.118889 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.118913 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.118930 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.221529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.221585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.221599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.221620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.221636 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.323885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.323944 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.323964 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.323986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.324003 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.426378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.426437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.426453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.426475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.426490 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.528771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.528824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.528836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.528857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.528869 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.632481 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.632557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.632581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.632611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.632634 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.678493 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.678515 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.678536 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:45 crc kubenswrapper[4926]: E1007 20:56:45.678629 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:45 crc kubenswrapper[4926]: E1007 20:56:45.678772 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:45 crc kubenswrapper[4926]: E1007 20:56:45.678890 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.735668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.735728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.735746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.735772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.735788 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.838320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.838368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.838387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.838409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.838426 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.941902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.941963 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.941981 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.942006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:45 crc kubenswrapper[4926]: I1007 20:56:45.942026 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:45Z","lastTransitionTime":"2025-10-07T20:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.045472 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.045529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.045552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.045581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.045603 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.148249 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.148324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.148345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.148369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.148386 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.251153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.251251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.251280 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.251319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.251330 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.353125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.353157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.353166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.353183 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.353214 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.455537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.455571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.455582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.455598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.455608 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.558279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.558332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.558346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.558368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.558383 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.662363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.662488 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.662520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.662555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.662587 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.678135 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:46 crc kubenswrapper[4926]: E1007 20:56:46.678404 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.766277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.766349 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.766368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.766397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.766418 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.870163 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.870226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.870237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.870252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.870263 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.973509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.973574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.973592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.973625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:46 crc kubenswrapper[4926]: I1007 20:56:46.973643 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:46Z","lastTransitionTime":"2025-10-07T20:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.075745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.075810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.075827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.075850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.075867 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.178797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.178862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.178871 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.178886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.178896 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.281796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.281854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.281871 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.281895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.281914 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.384819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.384881 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.384898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.384921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.384938 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.487265 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.487344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.487370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.487400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.487419 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.590476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.590548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.590566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.590593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.590610 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.678259 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.678383 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.678267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:47 crc kubenswrapper[4926]: E1007 20:56:47.678479 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:47 crc kubenswrapper[4926]: E1007 20:56:47.678593 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:47 crc kubenswrapper[4926]: E1007 20:56:47.678757 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.693724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.693780 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.693794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.693814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.694164 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.797233 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.797296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.797317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.797342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.797359 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.900837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.900895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.900913 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.900936 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.900954 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.932707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.932777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.932801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.932829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.932860 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: E1007 20:56:47.954547 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:47Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.959852 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.959901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.959918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.959942 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.959959 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:47 crc kubenswrapper[4926]: E1007 20:56:47.980421 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:47Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.985048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.985232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.985261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.985292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:47 crc kubenswrapper[4926]: I1007 20:56:47.985318 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:47Z","lastTransitionTime":"2025-10-07T20:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: E1007 20:56:48.006016 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:48Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.010671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.010705 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.010719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.010749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.010764 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: E1007 20:56:48.029918 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:48Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.034612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.034665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.034688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.034715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.034739 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: E1007 20:56:48.054667 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:48Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:48 crc kubenswrapper[4926]: E1007 20:56:48.054908 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.057133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.057239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.057267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.057298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.057321 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.159757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.159825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.159848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.159878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.159901 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.263351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.263425 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.263448 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.263475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.263491 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.367080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.367138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.367155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.367177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.367226 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.470286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.470355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.470377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.470407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.470431 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.573395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.573458 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.573481 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.573509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.573529 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.676629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.676722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.676747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.676779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.676805 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.678327 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:48 crc kubenswrapper[4926]: E1007 20:56:48.678483 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.780059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.780104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.780122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.780144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.780161 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.883170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.883241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.883254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.883273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.883286 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.986803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.986862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.986880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.986904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:48 crc kubenswrapper[4926]: I1007 20:56:48.986920 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:48Z","lastTransitionTime":"2025-10-07T20:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.089159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.089261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.089286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.089314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.089337 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.192962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.193024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.193040 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.193066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.193085 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.296149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.296251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.296277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.296311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.296333 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.398633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.398698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.398723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.398753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.398778 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.501462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.501531 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.501553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.501583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.501608 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.605808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.605887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.605909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.605941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.605963 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.678825 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.678889 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.678850 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:49 crc kubenswrapper[4926]: E1007 20:56:49.679069 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:49 crc kubenswrapper[4926]: E1007 20:56:49.679145 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:49 crc kubenswrapper[4926]: E1007 20:56:49.679270 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.709378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.709450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.709467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.709491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.709512 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.813460 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.813561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.813584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.813616 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.813639 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.916589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.916644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.916663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.916689 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:49 crc kubenswrapper[4926]: I1007 20:56:49.916707 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:49Z","lastTransitionTime":"2025-10-07T20:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.019567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.019629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.019647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.019678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.019695 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.122764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.122827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.122844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.122869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.122886 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.226777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.226869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.226894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.226924 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.226948 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.329845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.329914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.329937 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.329971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.329994 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.433879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.433977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.433998 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.434023 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.434040 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.537399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.537507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.537524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.537550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.537568 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.640062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.640236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.640262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.640293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.640315 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.677949 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:50 crc kubenswrapper[4926]: E1007 20:56:50.678142 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.697164 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.715508 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.735553 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.743383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.743433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.743450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.743474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.743492 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.758481 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.779548 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.801044 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.818954 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.832283 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.845462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.845509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.845520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.845536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.845548 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.875007 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.896388 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.914765 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.930932 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.947967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.948272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.948466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.948625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.948753 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:50Z","lastTransitionTime":"2025-10-07T20:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.957127 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.978275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:50 crc kubenswrapper[4926]: I1007 20:56:50.998450 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:50Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.016896 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:51Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.036460 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:51Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.051311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.051375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.051394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.051424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.051442 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.070312 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:51Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.153751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.153858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.153876 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.153898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.153915 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.257071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.257120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.257138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.257161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.257177 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.360155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.360578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.360597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.360623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.360641 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.464295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.464355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.464375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.464400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.464417 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.566508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.566546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.566555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.566568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.566576 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.669493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.669570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.669597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.669651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.669676 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.677906 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.677921 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:51 crc kubenswrapper[4926]: E1007 20:56:51.678121 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.677921 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:51 crc kubenswrapper[4926]: E1007 20:56:51.678301 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:51 crc kubenswrapper[4926]: E1007 20:56:51.678441 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.772825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.772894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.772931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.772961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.772985 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.876370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.876436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.876453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.876477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.876495 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.978436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.978667 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.978764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.978843 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:51 crc kubenswrapper[4926]: I1007 20:56:51.978910 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:51Z","lastTransitionTime":"2025-10-07T20:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.082003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.082036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.082046 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.082060 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.082069 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.185713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.185765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.185779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.185797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.185816 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.289450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.289510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.289529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.289554 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.289573 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.392035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.392075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.392085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.392103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.392114 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.494632 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.494675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.494684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.494697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.494706 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.600081 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.600140 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.600156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.600179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.600220 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.678054 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:52 crc kubenswrapper[4926]: E1007 20:56:52.678252 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.679580 4926 scope.go:117] "RemoveContainer" containerID="7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d" Oct 07 20:56:52 crc kubenswrapper[4926]: E1007 20:56:52.680014 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.703286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.703350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.703377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.703407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.703434 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.806561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.806612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.806628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.806674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.806689 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.909479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.909543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.909560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.909581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:52 crc kubenswrapper[4926]: I1007 20:56:52.909596 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:52Z","lastTransitionTime":"2025-10-07T20:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.011340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.011383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.011396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.011415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.011430 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.113029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.113067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.113077 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.113091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.113102 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.216679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.216734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.216751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.216776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.216793 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.319719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.319775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.319792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.319815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.319832 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.423108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.423229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.423255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.423314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.423342 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.526791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.526850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.526870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.526901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.526918 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.630270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.630337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.630355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.630379 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.630396 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.677823 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.677855 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.677955 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:53 crc kubenswrapper[4926]: E1007 20:56:53.678055 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:53 crc kubenswrapper[4926]: E1007 20:56:53.678151 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:53 crc kubenswrapper[4926]: E1007 20:56:53.678314 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.733814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.733853 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.733865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.733881 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.733893 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.836181 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.836248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.836261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.836279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.836291 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.939013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.939069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.939086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.939121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:53 crc kubenswrapper[4926]: I1007 20:56:53.939142 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:53Z","lastTransitionTime":"2025-10-07T20:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.041160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.041220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.041236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.041253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.041266 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.143775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.143811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.143823 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.143838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.143849 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.246647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.246710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.246774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.246803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.246832 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.349779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.349838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.349856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.349879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.349897 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.452594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.452633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.452646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.452663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.452731 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.556101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.556159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.556182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.556304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.556322 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.659444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.659479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.659492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.659514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.659528 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.679050 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:54 crc kubenswrapper[4926]: E1007 20:56:54.679300 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.761671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.761721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.761734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.761751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.761764 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.864442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.864491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.864503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.864520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.864531 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.966658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.966713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.966732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.966755 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:54 crc kubenswrapper[4926]: I1007 20:56:54.966772 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:54Z","lastTransitionTime":"2025-10-07T20:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.069504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.069541 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.069550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.069602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.069614 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.172528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.172610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.172637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.172675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.172699 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.275357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.275416 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.275433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.275457 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.275475 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.378327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.378384 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.378401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.378426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.378443 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.481212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.481253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.481261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.481276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.481287 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.584084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.584232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.584253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.584278 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.584301 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.678104 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.678181 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.678265 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:55 crc kubenswrapper[4926]: E1007 20:56:55.678442 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:55 crc kubenswrapper[4926]: E1007 20:56:55.678561 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:55 crc kubenswrapper[4926]: E1007 20:56:55.678746 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.687311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.687359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.687376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.687400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.687417 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.789937 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.789989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.790006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.790029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.790046 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.892947 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.892994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.893010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.893032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.893049 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.995670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.995734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.995752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.995776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:55 crc kubenswrapper[4926]: I1007 20:56:55.995793 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:55Z","lastTransitionTime":"2025-10-07T20:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.097967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.098026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.098043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.098067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.098085 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.200480 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.200537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.200558 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.200587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.200609 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.302642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.302702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.302718 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.302754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.302772 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.405563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.405626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.405644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.405668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.405683 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.508818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.508862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.508879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.508900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.508913 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.612560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.612629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.612683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.612715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.612735 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.679035 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:56 crc kubenswrapper[4926]: E1007 20:56:56.679374 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.701923 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.716131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.716214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.716231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.716253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.716275 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.819071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.819136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.819154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.819179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.819228 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.921951 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.922029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.922051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.922080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:56 crc kubenswrapper[4926]: I1007 20:56:56.922104 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:56Z","lastTransitionTime":"2025-10-07T20:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.024590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.024647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.024665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.024683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.024696 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.127262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.127321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.127335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.127365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.127379 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.230529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.230604 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.230629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.230672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.230697 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.334536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.334614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.334640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.334670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.334694 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.437164 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.437219 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.437228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.437241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.437249 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.541072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.541133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.541177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.541232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.541256 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.644007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.644056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.644073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.644095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.644112 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.678683 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.678684 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:57 crc kubenswrapper[4926]: E1007 20:56:57.678959 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:57 crc kubenswrapper[4926]: E1007 20:56:57.678789 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.678692 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:57 crc kubenswrapper[4926]: E1007 20:56:57.679082 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.746023 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.746067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.746080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.746096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.746109 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.848321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.848361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.848373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.848393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.848410 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.951162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.951213 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.951225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.951242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:57 crc kubenswrapper[4926]: I1007 20:56:57.951252 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:57Z","lastTransitionTime":"2025-10-07T20:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.053391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.053461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.053478 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.053500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.053517 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.156070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.156129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.156151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.156178 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.156229 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.258762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.258847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.258865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.258886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.258901 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.348700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.348763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.348779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.348800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.348817 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.364511 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:58Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.368746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.368801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.368817 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.368839 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.368854 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.383713 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:58Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.388687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.388792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.388816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.388888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.388912 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.405337 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:58Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.409496 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.409551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.409565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.409581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.409593 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.426860 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:58Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.430493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.430523 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.430533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.430550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.430564 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.445978 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:56:58Z is after 2025-08-24T17:21:41Z" Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.446138 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.447180 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.447432 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.447543 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:57:30.447508113 +0000 UTC m=+100.486089303 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.447581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.447648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.447666 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.447716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.447733 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.550851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.550924 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.550947 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.550980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.551003 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.652744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.652803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.652826 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.652851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.652870 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.678408 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:56:58 crc kubenswrapper[4926]: E1007 20:56:58.678555 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.755975 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.756027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.756039 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.756059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.756071 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.858945 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.859004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.859015 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.859029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.859038 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.961409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.961673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.961836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.962034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:58 crc kubenswrapper[4926]: I1007 20:56:58.962230 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:58Z","lastTransitionTime":"2025-10-07T20:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.064390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.064481 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.064501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.064524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.064542 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.167397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.167453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.167470 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.167495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.167512 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.270769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.270844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.270865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.270893 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.270912 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.373501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.373563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.373582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.373607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.373624 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.476904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.477001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.477018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.477071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.477096 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.579318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.579357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.579366 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.579382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.579396 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.677853 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:56:59 crc kubenswrapper[4926]: E1007 20:56:59.677990 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.677876 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:56:59 crc kubenswrapper[4926]: E1007 20:56:59.678094 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.677856 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:56:59 crc kubenswrapper[4926]: E1007 20:56:59.678162 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.681117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.681168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.681179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.681211 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.681221 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.783529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.783573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.783582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.783598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.783608 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.885912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.885947 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.885956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.885969 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.885978 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.988865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.988905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.988917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.988932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:56:59 crc kubenswrapper[4926]: I1007 20:56:59.988944 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:56:59Z","lastTransitionTime":"2025-10-07T20:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.091486 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.091513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.091521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.091536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.091545 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.114824 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/0.log" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.114870 4926 generic.go:334] "Generic (PLEG): container finished" podID="b1217995-75cc-4fc7-8f5f-7752d2e378e2" containerID="8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6" exitCode=1 Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.114899 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerDied","Data":"8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.115349 4926 scope.go:117] "RemoveContainer" containerID="8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.147234 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.165350 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.182063 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.193957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.193992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.194004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.194027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.194038 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.199156 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.215250 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.230950 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.248174 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.266167 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.280886 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.296855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.296894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.296905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.296922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.296931 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.303927 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.316061 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.327918 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.338128 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.353929 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.363904 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.374678 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.389358 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.398937 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.398968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.398979 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.398995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.399006 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.403855 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.413333 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.501160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.501251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.501269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.501293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.501312 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.604283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.604348 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.604372 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.604406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.604431 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.678430 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:00 crc kubenswrapper[4926]: E1007 20:57:00.678678 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.697766 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.706966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.707028 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.707050 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.707078 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.707100 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.710034 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.721113 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.735394 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.747759 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.761093 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.772486 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.784538 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.794248 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.802973 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.808801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.808827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.808835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.808849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.808858 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.811633 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.828343 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.838110 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.851369 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.864062 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.878992 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.892062 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.907280 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.910765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.910791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.910799 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.910814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.910823 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:00Z","lastTransitionTime":"2025-10-07T20:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:00 crc kubenswrapper[4926]: I1007 20:57:00.932177 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:00Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.013168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.013392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.013452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.013543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.013613 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.116702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.116741 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.116753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.116769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.116796 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.119828 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/0.log" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.119880 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerStarted","Data":"31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.137816 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.154743 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.176707 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.189221 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.202749 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.214565 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.218677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.218698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.218707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.218721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.218731 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.229026 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.239320 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.249685 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.266834 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.280443 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.296035 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.306648 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.319726 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.321101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.321141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.321154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.321172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.321186 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.330666 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.339826 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.351007 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.361653 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.383585 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:01Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.423317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.423370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.423386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.423409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.423426 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.525493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.525537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.525555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.525576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.525594 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.628688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.628774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.628793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.628819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.628840 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.678643 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.678650 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.678648 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:01 crc kubenswrapper[4926]: E1007 20:57:01.678784 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:01 crc kubenswrapper[4926]: E1007 20:57:01.679095 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:01 crc kubenswrapper[4926]: E1007 20:57:01.679270 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.731899 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.731941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.731953 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.731970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.731982 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.834668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.834763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.834781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.834814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.834834 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.937143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.937207 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.937220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.937236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:01 crc kubenswrapper[4926]: I1007 20:57:01.937247 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:01Z","lastTransitionTime":"2025-10-07T20:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.039964 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.040004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.040023 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.040047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.040070 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.142477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.142514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.142526 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.142543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.142555 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.245898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.245957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.245977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.245999 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.246014 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.349426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.349484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.349505 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.349530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.349562 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.456627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.456692 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.456710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.456733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.456749 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.558608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.558659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.558676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.558699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.558716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.660737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.660786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.660795 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.660811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.660822 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.678324 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:02 crc kubenswrapper[4926]: E1007 20:57:02.678495 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.763549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.763603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.763620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.763644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.763661 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.866035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.866092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.866109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.866133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.866149 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.969183 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.969260 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.969275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.969299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:02 crc kubenswrapper[4926]: I1007 20:57:02.969318 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:02Z","lastTransitionTime":"2025-10-07T20:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.071927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.071984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.072001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.072024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.072043 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.174973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.175033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.175051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.175075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.175093 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.277099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.277141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.277149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.277162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.277174 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.379719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.379752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.379761 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.379774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.379782 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.482844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.482904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.482923 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.483050 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.483074 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.585952 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.586026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.586049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.586084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.586107 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.678223 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.678260 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.678258 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:03 crc kubenswrapper[4926]: E1007 20:57:03.678348 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:03 crc kubenswrapper[4926]: E1007 20:57:03.678427 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:03 crc kubenswrapper[4926]: E1007 20:57:03.678564 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.689337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.689381 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.689398 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.689456 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.689471 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.792714 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.792771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.792789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.792813 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.792830 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.895561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.895621 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.895638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.895663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.895684 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.998026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.998085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.998103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.998126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:03 crc kubenswrapper[4926]: I1007 20:57:03.998144 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:03Z","lastTransitionTime":"2025-10-07T20:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.100526 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.100576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.100591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.100635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.100653 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.202803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.202838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.202851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.202867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.202879 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.305713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.305767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.305789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.305816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.305836 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.409268 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.409321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.409338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.409362 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.409379 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.511396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.511474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.511501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.511535 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.511561 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.615080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.615134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.615149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.615167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.615188 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.678507 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:04 crc kubenswrapper[4926]: E1007 20:57:04.678743 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.717963 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.718024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.718041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.718066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.718085 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.820883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.820944 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.820960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.820983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.821028 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.924018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.924066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.924087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.924116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:04 crc kubenswrapper[4926]: I1007 20:57:04.924137 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:04Z","lastTransitionTime":"2025-10-07T20:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.026244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.026314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.026326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.026343 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.026354 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.128696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.128759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.128778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.128809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.128832 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.232160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.232264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.232290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.232318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.232335 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.335812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.335867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.335883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.335907 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.335923 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.439163 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.439243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.439261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.439284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.439302 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.542445 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.542524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.542547 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.542578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.542600 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.645768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.645849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.645866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.645892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.645910 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.678813 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.678870 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.678886 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:05 crc kubenswrapper[4926]: E1007 20:57:05.679010 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:05 crc kubenswrapper[4926]: E1007 20:57:05.679160 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:05 crc kubenswrapper[4926]: E1007 20:57:05.679305 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.749234 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.749294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.749313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.749336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.749353 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.852855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.852938 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.852962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.852993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.853018 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.956149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.956256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.956279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.956309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:05 crc kubenswrapper[4926]: I1007 20:57:05.956332 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:05Z","lastTransitionTime":"2025-10-07T20:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.059536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.059602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.059625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.059652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.059671 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.161570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.161613 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.161622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.161637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.161648 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.264490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.264546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.264565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.264589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.264608 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.367327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.367395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.367413 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.367439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.367460 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.470706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.470786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.470804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.470828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.470845 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.574241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.574296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.574310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.574332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.574351 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.676467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.676506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.676514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.676526 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.676534 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.679364 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:06 crc kubenswrapper[4926]: E1007 20:57:06.679931 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.680258 4926 scope.go:117] "RemoveContainer" containerID="7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.778500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.778534 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.778545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.778561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.778571 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.880550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.880582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.880593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.880610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.880620 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.983301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.983344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.983354 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.983368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:06 crc kubenswrapper[4926]: I1007 20:57:06.983380 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:06Z","lastTransitionTime":"2025-10-07T20:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.086155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.086241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.086250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.086263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.086272 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.139777 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/2.log" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.142869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.143422 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.163032 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.185164 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.188649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.188690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.188708 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.188728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.188739 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.203408 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.215627 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.228395 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.247424 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.260017 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.277274 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.292065 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.292140 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.292160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.292184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.292224 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.292272 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.310252 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.325251 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.338257 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.356576 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.371293 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.384587 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.395021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.395057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.395067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.395082 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.395092 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.405176 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.422453 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.436242 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.451902 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:07Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.497730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.497784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.497802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.497824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.497841 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.601186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.601282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.601299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.601323 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.601343 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.677949 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.678029 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:07 crc kubenswrapper[4926]: E1007 20:57:07.678075 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.678138 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:07 crc kubenswrapper[4926]: E1007 20:57:07.678251 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:07 crc kubenswrapper[4926]: E1007 20:57:07.678381 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.702968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.703740 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.703892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.704034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.704159 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.807994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.808068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.808090 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.808115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.808134 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.911910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.911948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.911959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.911973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:07 crc kubenswrapper[4926]: I1007 20:57:07.911983 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:07Z","lastTransitionTime":"2025-10-07T20:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.015166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.015188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.015212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.015226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.015234 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.118173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.118273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.118296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.118329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.118387 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.148685 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/3.log" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.149276 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/2.log" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.152779 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" exitCode=1 Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.152831 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.152881 4926 scope.go:117] "RemoveContainer" containerID="7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.153261 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.153391 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.176261 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.201022 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.217048 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.221356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.221395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.221415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.221439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.221502 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.236896 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.261922 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.279955 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.298050 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.313998 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.324140 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.324228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.324247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.324274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.324291 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.332865 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.366404 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.387318 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.409804 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.427537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.427588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.427606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.427630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.427646 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.427781 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.450716 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.470692 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.488879 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.506722 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.528697 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.530824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.530900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.530918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.530945 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.530992 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.560001 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4f291cae0372d5bfae39f6a82ac2ffad3c20a47d432ebdeff146440579295d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:37Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.573673 6590 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.573829 6590 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574123 6590 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 20:56:37.574313 6590 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1007 20:56:37.574805 6590 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 20:56:37.574878 6590 factory.go:656] Stopping watch factory\\\\nI1007 20:56:37.574918 6590 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 20:56:37.613875 6590 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 20:56:37.613916 6590 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 20:56:37.614040 6590 ovnkube.go:599] Stopped ovnkube\\\\nI1007 20:56:37.614074 6590 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 20:56:37.614228 6590 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:57:07Z\\\",\\\"message\\\":\\\"D:\\\\\\\"cab7c637-a021-4a4d-a4b9-06d63c44316f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1007 20:57:07.621437 6943 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInfor\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.634031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.634108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.634129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.634156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.634175 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.679081 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.679346 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.703657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.703727 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.703745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.703767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.703785 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.724586 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.730700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.730759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.730776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.730800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.730817 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.752842 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.757801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.757933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.757995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.758025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.758075 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.780116 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.785709 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.785766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.785784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.785810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.785829 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.804984 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.810497 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.810552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.810570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.810594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.810611 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.830180 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:08Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:08 crc kubenswrapper[4926]: E1007 20:57:08.830485 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.832738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.832801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.832822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.832852 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.832875 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.935928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.935996 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.936020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.936051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:08 crc kubenswrapper[4926]: I1007 20:57:08.936072 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:08Z","lastTransitionTime":"2025-10-07T20:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.038463 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.038542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.038559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.038572 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.038582 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.141513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.141571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.141592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.141621 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.141651 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.160059 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/3.log" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.165320 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 20:57:09 crc kubenswrapper[4926]: E1007 20:57:09.165733 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.188591 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.208478 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.225988 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.245013 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.245267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.245301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.245321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.245343 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.245360 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.264872 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.299909 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:57:07Z\\\",\\\"message\\\":\\\"D:\\\\\\\"cab7c637-a021-4a4d-a4b9-06d63c44316f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1007 20:57:07.621437 6943 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInfor\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:57:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.314939 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.329796 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.348658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.348749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.348768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.348791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.348808 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.350169 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.375517 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.391595 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.410491 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.429471 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.444333 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.451873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.451967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.452031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.452058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.452090 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.460868 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.492745 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.513130 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.531649 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.551852 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:09Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.555130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.555255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.555281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.555309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.555331 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.658991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.659055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.659072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.659097 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.659114 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.678436 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.678502 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.678462 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:09 crc kubenswrapper[4926]: E1007 20:57:09.678665 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:09 crc kubenswrapper[4926]: E1007 20:57:09.678813 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:09 crc kubenswrapper[4926]: E1007 20:57:09.678981 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.761803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.761881 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.761901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.761929 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.761954 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.864995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.865062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.865085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.865116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.865145 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.968138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.968231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.968250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.968274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:09 crc kubenswrapper[4926]: I1007 20:57:09.968292 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:09Z","lastTransitionTime":"2025-10-07T20:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.071373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.071463 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.071485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.071517 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.071541 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.174306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.174421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.174442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.174466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.174484 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.277654 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.277724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.277742 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.277768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.277790 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.380888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.380938 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.380951 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.380967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.380981 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.484497 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.484562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.484583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.484609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.484626 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.587579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.587651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.587675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.587703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.587727 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.678380 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:10 crc kubenswrapper[4926]: E1007 20:57:10.678572 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.690576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.690630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.690651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.690681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.690702 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.694377 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.713895 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.735305 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.749152 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.766184 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.792600 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.793470 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.793531 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.793548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.793574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.793591 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.810139 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.824940 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.840612 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.874825 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.896284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.896342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.896359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.896383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.896401 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.899096 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.917711 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.937811 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.960939 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.984640 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:10Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:10 crc kubenswrapper[4926]: I1007 20:57:10.999809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:10.999858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:10.999875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:10.999898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:10.999915 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:10Z","lastTransitionTime":"2025-10-07T20:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.003147 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.022151 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.044792 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.076171 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:57:07Z\\\",\\\"message\\\":\\\"D:\\\\\\\"cab7c637-a021-4a4d-a4b9-06d63c44316f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1007 20:57:07.621437 6943 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInfor\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:57:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:11Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.102661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.102730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.102747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.102772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.102789 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.205644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.205691 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.205707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.205730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.205748 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.308990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.309052 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.309068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.309092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.309112 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.412533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.412594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.412612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.412640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.412657 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.515277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.515346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.515373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.515405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.515427 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.618374 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.618425 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.618442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.618465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.618481 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.678792 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.678876 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:11 crc kubenswrapper[4926]: E1007 20:57:11.679006 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.678791 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:11 crc kubenswrapper[4926]: E1007 20:57:11.679169 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:11 crc kubenswrapper[4926]: E1007 20:57:11.679346 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.721702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.721767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.721785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.721809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.721828 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.824518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.824575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.824592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.824622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.824662 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.927772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.927851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.927873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.927904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:11 crc kubenswrapper[4926]: I1007 20:57:11.927925 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:11Z","lastTransitionTime":"2025-10-07T20:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.031305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.031385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.031409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.031442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.031466 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.139844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.139910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.139929 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.139955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.139974 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.242870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.242928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.242946 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.242970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.242989 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.345762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.345807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.345825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.345846 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.345860 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.449594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.449652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.449671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.449697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.449715 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.552836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.552900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.552917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.552944 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.553189 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.656387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.656450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.656468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.656491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.656510 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.678567 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:12 crc kubenswrapper[4926]: E1007 20:57:12.678885 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.759500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.760034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.760185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.760386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.760575 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.863988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.864058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.864075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.864104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.864120 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.967021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.967111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.967125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.967159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:12 crc kubenswrapper[4926]: I1007 20:57:12.967177 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:12Z","lastTransitionTime":"2025-10-07T20:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.070242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.070302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.070325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.070357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.070381 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.173143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.173267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.173304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.173334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.173358 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.277106 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.277593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.277611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.277636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.277652 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.380625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.380684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.380701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.380726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.380743 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.483449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.483515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.483538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.483568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.483592 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.587321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.587399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.587421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.587450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.587473 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.678507 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.678536 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.678553 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:13 crc kubenswrapper[4926]: E1007 20:57:13.678657 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:13 crc kubenswrapper[4926]: E1007 20:57:13.678810 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:13 crc kubenswrapper[4926]: E1007 20:57:13.678874 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.691116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.691181 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.691257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.691286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.691308 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.794338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.794396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.794416 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.794443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.794460 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.897649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.897726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.897749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.897781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:13 crc kubenswrapper[4926]: I1007 20:57:13.897803 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:13Z","lastTransitionTime":"2025-10-07T20:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.000396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.000469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.000491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.000521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.000542 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.104250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.104286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.104331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.104345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.104353 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.206984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.207036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.207059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.207092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.207117 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.310646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.310712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.310734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.310764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.310782 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.413637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.413704 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.413723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.413749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.413766 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.448527 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.448724 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.44869866 +0000 UTC m=+148.487279850 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.516805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.516887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.516904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.516925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.516942 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.549406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.549494 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549521 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549601 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.549577626 +0000 UTC m=+148.588158786 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.549532 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549659 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.549672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549729 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.54970743 +0000 UTC m=+148.588288620 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549822 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549827 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549937 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549969 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.549839 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.550060 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.550064 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.550031289 +0000 UTC m=+148.588612539 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.550148 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.550124302 +0000 UTC m=+148.588705582 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.620225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.620282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.620305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.620335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.620361 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.678687 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:14 crc kubenswrapper[4926]: E1007 20:57:14.678853 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.723581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.723683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.723720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.723749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.723770 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.827390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.827450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.827468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.827493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.827509 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.930388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.930446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.930466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.930497 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:14 crc kubenswrapper[4926]: I1007 20:57:14.930520 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:14Z","lastTransitionTime":"2025-10-07T20:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.033960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.034030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.034049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.034071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.034089 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.137814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.137869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.137889 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.137914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.137934 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.241973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.242043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.242060 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.242086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.242104 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.345053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.345136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.345158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.345186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.345243 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.448106 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.448161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.448184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.448243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.448262 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.551948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.552029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.552047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.552073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.552091 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.655510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.655592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.655611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.655637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.655655 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.678573 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.678590 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:15 crc kubenswrapper[4926]: E1007 20:57:15.678756 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:15 crc kubenswrapper[4926]: E1007 20:57:15.678886 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.678607 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:15 crc kubenswrapper[4926]: E1007 20:57:15.679132 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.758573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.758636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.758653 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.758677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.758696 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.861735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.861777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.861785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.861801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.861811 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.965072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.965331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.965345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.965360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:15 crc kubenswrapper[4926]: I1007 20:57:15.965370 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:15Z","lastTransitionTime":"2025-10-07T20:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.067932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.067982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.068003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.068031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.068050 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.170614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.170685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.170708 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.170737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.170758 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.273980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.274056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.274080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.274144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.274167 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.377060 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.377140 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.377159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.377185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.377232 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.480037 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.480088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.480099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.480114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.480125 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.582742 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.582789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.582804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.582823 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.582835 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.678367 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:16 crc kubenswrapper[4926]: E1007 20:57:16.678530 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.685119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.685156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.685166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.685223 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.685234 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.787725 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.787802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.787833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.787857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.787870 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.891109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.891164 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.891180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.891240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.891264 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.994751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.994835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.994854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.994877 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:16 crc kubenswrapper[4926]: I1007 20:57:16.995684 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:16Z","lastTransitionTime":"2025-10-07T20:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.098636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.098739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.098765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.098796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.098822 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.201431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.201479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.201491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.201508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.201525 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.304337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.304412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.304427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.304448 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.304462 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.407415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.407476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.407551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.407597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.407621 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.510683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.510776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.510801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.511375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.511431 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.614853 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.614916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.614933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.614958 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.614975 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.678045 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.678248 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.678452 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:17 crc kubenswrapper[4926]: E1007 20:57:17.678670 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:17 crc kubenswrapper[4926]: E1007 20:57:17.678809 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:17 crc kubenswrapper[4926]: E1007 20:57:17.679016 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.718244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.718291 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.718304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.718324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.718336 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.821816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.821878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.821898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.821923 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.821942 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.925123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.925240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.925266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.925299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:17 crc kubenswrapper[4926]: I1007 20:57:17.925323 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:17Z","lastTransitionTime":"2025-10-07T20:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.028811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.028879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.028903 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.028933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.028956 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.131698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.131760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.131777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.131803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.131823 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.234636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.234728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.234751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.234778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.234795 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.338475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.338724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.338874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.339007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.339161 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.442086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.442144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.442165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.442189 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.442246 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.544926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.544995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.545017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.545044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.545062 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.648288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.648365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.648393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.648425 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.648448 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.678560 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:18 crc kubenswrapper[4926]: E1007 20:57:18.678792 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.751171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.751257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.751275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.751299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.751316 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.853490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.853545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.853569 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.853590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.853605 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.957085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.957150 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.957173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.957247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:18 crc kubenswrapper[4926]: I1007 20:57:18.957278 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:18Z","lastTransitionTime":"2025-10-07T20:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.059870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.059910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.059920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.059936 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.059948 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.083934 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.084006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.084024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.084047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.084065 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.103618 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.107873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.107933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.107950 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.107983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.108033 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.122090 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.127045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.127118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.127136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.127162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.127180 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.144114 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.149249 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.149321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.149339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.149364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.149383 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.168532 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.175299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.175851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.175878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.175898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.175912 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.193508 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:19Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.193670 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.195524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.195598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.195624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.195657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.195681 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.298105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.298137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.298145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.298175 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.298183 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.400551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.400608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.400625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.400649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.400664 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.503868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.503942 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.503966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.503996 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.504016 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.606980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.607016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.607033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.607055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.607068 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.678313 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.678400 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.678440 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.678621 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.678629 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:19 crc kubenswrapper[4926]: E1007 20:57:19.678667 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.709810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.709856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.709867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.709885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.709897 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.812266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.812308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.812320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.812337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.812349 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.935929 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.935985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.936001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.936024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:19 crc kubenswrapper[4926]: I1007 20:57:19.936043 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:19Z","lastTransitionTime":"2025-10-07T20:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.039047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.039121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.039141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.039166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.039183 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.142750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.142816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.142855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.142891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.142914 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.245487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.245562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.245587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.245617 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.245662 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.348068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.348406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.348415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.348427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.348435 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.450812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.450870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.450887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.450909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.450925 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.554082 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.554132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.554143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.554161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.554172 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.657365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.657449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.657484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.657514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.657535 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.677963 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:20 crc kubenswrapper[4926]: E1007 20:57:20.678184 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.699339 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"809bf70c-c24b-4406-a428-96c0b21a120f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d2650887e67a6ee84341d7e57a28eb16ea2dec772ddeaaeb0f6adfb6490e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://240046ebb1b3acad60177ca6c156b729aac5aacc22671f7b0a4bce6d0a0cad8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70d631c9731113c9f22babf8c78430fe528190aa40f790dbf0769b1e47db188\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f22ca106ef2565530b88dad0f9dfd2718084d632ff76678166dc8557473524e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72269e1b7b911708c3e51a547d5891de03b77735ae2fe2c6c27519551f76def7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1007 20:56:04.925814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 20:56:04.929492 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753227785/tls.crt::/tmp/serving-cert-3753227785/tls.key\\\\\\\"\\\\nI1007 20:56:10.545722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 20:56:10.550332 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 20:56:10.550365 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 20:56:10.550409 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 20:56:10.550420 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 20:56:10.560185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1007 20:56:10.560239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560246 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 20:56:10.560252 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 20:56:10.560257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 20:56:10.560261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 20:56:10.560265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 20:56:10.560287 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1007 20:56:10.563418 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db13a8c6aadfe53f9f8f9cb6dd26680756f81c7dd154f336ae1674f1f342f18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d8094834f4cdaa84bcad53d62e84f072010f4cdb36c1c6acab2b2931fc6140\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.717018 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57eb78b0-3098-4fd6-92e2-0753dddcbc59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b709bf826ee7d86fd167499a569099abe64c956f8bca5972bd888a6ce975b93b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd9f0b82b36bd4e29b824d37a06c4704798fbdbe0534850276eb5e02be6bb68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0879d4a1f96f2dc995135affb661749505c06ab6cafd4c2cce5dce8cfeea9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7980e7f2ad08e304553ed302034c5ebd419e868d9edcf7015063d034998dc3d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.736664 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d414b86-517f-4a92-8cce-ae4dd7b3c2e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e518e3f9f0265ab31aedeaeafd179d20463b3042033c3caec1980d887c167fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0273412d9631b917d0f72ef97c2f2beaf0dc109542af5fdd44159d59e5b498fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8d341b7e2290ca429363b8dd30b7be72280ddc3142f5cbdf860868780c8bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51f748c5c11f8b8a745c18909b36b5f979a90683f4e467bbdb3fb5f3a800feac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.752721 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.760977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.761102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.761124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.761151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.761171 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.770352 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bec39aa071f4308c2f5582b9927574e4e380a1a06c94b436c24857b38c9f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.802054 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8c34cd-2651-4f25-8439-1897f0e937b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:57:07Z\\\",\\\"message\\\":\\\"D:\\\\\\\"cab7c637-a021-4a4d-a4b9-06d63c44316f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1007 20:57:07.621437 6943 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInfor\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:57:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc8t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2zw6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.818984 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e78c3795-277b-4a59-90f5-8300ad92d53f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e3e2600cbc1ff0ad8ce1abe336766db50a11f07cc1995e9322927e673e7fa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://879dcbc8662a4bc8d090d8c9a70ab2f9222538216c08c25f8505fb94a98576f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.838555 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.857039 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.864664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.864743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.864767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.864796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.864818 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.874942 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8de1b64212285f6cad420ff70c15ce0ff11efea37018eddd8518843e1fcaf67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.890996 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rmrts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a38bebbe-fe67-479d-9790-9487f62944b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8dd7f212f01b299bffc15c19db1c77315bd7fe78bd0360f1db5f9ecc8788b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rfltx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rmrts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.906597 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4btw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f15f8f8-8603-4ecd-9035-3aeb61ab3d8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fda85663a225bb19aaccf40f2d27e09041a587c7a26ff20db8fc1c2f92187c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gm7md\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4btw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.926589 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2tlv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1217995-75cc-4fc7-8f5f-7752d2e378e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T20:56:59Z\\\",\\\"message\\\":\\\"2025-10-07T20:56:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb\\\\n2025-10-07T20:56:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2f561594-17c4-46f5-bf00-4236bf4e7afb to /host/opt/cni/bin/\\\\n2025-10-07T20:56:14Z [verbose] multus-daemon started\\\\n2025-10-07T20:56:14Z [verbose] Readiness Indicator file check\\\\n2025-10-07T20:56:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:57:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68x4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2tlv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.951197 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtr42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6eafe65-074f-4285-b66d-ce84d41ee711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b35d2e7ee6ea94a85c8b048efce84a121fa4d2e07aeac76f84993fb2109234d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7cf5c101cdf1f3c41bda3af1b854bd842bd1377577842c1b68977e8ba5f55bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9a4c4d30bd68bd494e35085077d3848d2a1e4736b352d194df098eb78b14b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b7b9c4c2819ea6ecefc0e7d342551b5d7f73442e24d6f677c8bf68d5b720fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://497d338045302619ce39db1b4a0d86742b484cc3d8a9efae126c8dcc009b1bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0937df5f1bda0b9e8fbd56625d7d2e18bdbf447afaf9df419dad251bb4b009\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11df518906c33a2af47cbec9aec463c15855b966141bdf413435f95b276ff552\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksxsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtr42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.968830 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.968875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.968886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.968904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.968916 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:20Z","lastTransitionTime":"2025-10-07T20:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:20 crc kubenswrapper[4926]: I1007 20:57:20.970335 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a8e664c-9b14-4510-8a51-bec46ecbaea5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j5g8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c6mr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:20Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.029004 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b51ad036-c93d-4a29-be6e-69d7103412a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5f5e9d4bfe021b5ae668b34b9ce8798fa3b5a53704f39ccbd52e81d5f70eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f829790f5eb44395c2be3ef7c4eea640e81d139bdfd00bf194fe4b1fe79ac7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://682898e5773e6d4774b74f90a1fd49facd4af270bfb62faa82f6cf495298498b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fecddd1871f288dab6f090a101552132246c7c7f6c0c0d3fcbe0a81ee4363caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a69f154945211b41cb6c7d2c66c7692ea3b6e7c96c27f2588d621c24b2c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf316b98de4faa203ca95208debe10c004793b74b4ef3e30dae9d824b4cb7ce9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d0fb58fb0f16abb34d608e1f20de6f2de7298cdc9561be55e220f4c4bb8541\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:52Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbd9133947349925ae42d2a3106dd9b0e1b82f133b13aa6ee429b0ccbc996dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T20:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:55:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:21Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.053005 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f653439010dca6c5ab127e3659e7b27eeef09fbff55bd9554dd8ac944490bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43459b4e8a3b3a2e2963f2ba48ff3c53e8e53651d35de0fed9251bbe592f0c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:21Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.070910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.070959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.070971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.070988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.071000 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.071571 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b0be3fadf126a29b3b4ee587db13da9a3b004cc8e12d92856c07413a870cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cb2vs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5cd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:21Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.083509 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cea4c9a-b58a-4d94-91ad-4056b318811a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T20:56:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8efa87da9bb4ae01fa600caae9e4fd9117d62ad153f60bec766da8d5391b03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc6138c4af41c4534f7a6c26217e9597e3dd3ea6b291391905c55386438d96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T20:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qzddl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T20:56:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6dbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:21Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.173673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.173747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.173813 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.173845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.173865 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.275634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.275673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.275685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.275702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.275715 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.378075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.378114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.378126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.378142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.378152 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.480867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.480916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.480928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.480946 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.480958 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.584012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.584071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.584088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.584111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.584128 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.678598 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:21 crc kubenswrapper[4926]: E1007 20:57:21.678759 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.678995 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:21 crc kubenswrapper[4926]: E1007 20:57:21.679119 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.679289 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:21 crc kubenswrapper[4926]: E1007 20:57:21.679457 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.686474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.686509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.686530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.686547 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.686556 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.789597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.789655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.789672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.789694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.789711 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.892996 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.893117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.893135 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.893156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.893173 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.996305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.996337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.996346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.996360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:21 crc kubenswrapper[4926]: I1007 20:57:21.996370 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:21Z","lastTransitionTime":"2025-10-07T20:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.098862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.098985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.099000 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.099018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.099029 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.202054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.202108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.202124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.202146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.202164 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.304552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.304599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.304610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.304625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.304637 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.407013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.407065 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.407075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.407096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.407111 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.509434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.509496 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.509515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.509540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.509560 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.612776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.612829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.612845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.612869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.612888 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.678368 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:22 crc kubenswrapper[4926]: E1007 20:57:22.678577 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.715427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.715503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.715525 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.715549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.715567 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.817744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.817804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.817821 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.817844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.817861 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.920822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.920868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.920878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.920892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:22 crc kubenswrapper[4926]: I1007 20:57:22.920901 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:22Z","lastTransitionTime":"2025-10-07T20:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.023840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.023882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.023892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.023907 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.023919 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.126246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.126279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.126288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.126303 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.126313 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.228768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.228798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.228807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.228819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.228827 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.331834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.331903 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.331924 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.331953 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.331976 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.435243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.435315 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.435338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.435366 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.435389 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.537556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.537612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.537628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.537651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.537669 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.640941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.641002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.641020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.641045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.641065 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.678056 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.678158 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.678167 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:23 crc kubenswrapper[4926]: E1007 20:57:23.678329 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:23 crc kubenswrapper[4926]: E1007 20:57:23.678640 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:23 crc kubenswrapper[4926]: E1007 20:57:23.678743 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.744513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.744601 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.744620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.744655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.744676 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.848266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.848326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.848345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.848369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.848387 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.951332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.951411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.951434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.951465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:23 crc kubenswrapper[4926]: I1007 20:57:23.951490 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:23Z","lastTransitionTime":"2025-10-07T20:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.055229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.055282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.055296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.055315 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.055331 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.159378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.159452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.159474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.159506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.159529 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.262435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.262481 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.262494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.262509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.262522 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.365872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.365933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.365951 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.365987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.366029 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.468925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.468983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.468994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.469008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.469017 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.571895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.571932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.571944 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.571960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.571971 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.674548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.674613 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.674632 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.674658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.674676 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.677946 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:24 crc kubenswrapper[4926]: E1007 20:57:24.678604 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.679078 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 20:57:24 crc kubenswrapper[4926]: E1007 20:57:24.679421 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.777748 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.777795 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.777808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.777826 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.777839 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.880781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.880816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.880825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.880842 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.880852 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.983793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.983845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.983863 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.983886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:24 crc kubenswrapper[4926]: I1007 20:57:24.983902 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:24Z","lastTransitionTime":"2025-10-07T20:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.086542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.086570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.086578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.086590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.086603 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.188900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.188956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.188973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.188995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.189014 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.293743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.293809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.293831 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.293860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.293888 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.396678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.396734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.396743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.396762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.396771 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.499180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.499281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.499306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.499335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.499356 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.602694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.602749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.602772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.602802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.602827 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.678706 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.678724 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:25 crc kubenswrapper[4926]: E1007 20:57:25.678904 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.678941 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:25 crc kubenswrapper[4926]: E1007 20:57:25.678999 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:25 crc kubenswrapper[4926]: E1007 20:57:25.679056 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.705961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.706016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.706029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.706048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.706060 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.814086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.814141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.814161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.814222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.814241 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.917392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.917457 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.917474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.917499 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:25 crc kubenswrapper[4926]: I1007 20:57:25.917516 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:25Z","lastTransitionTime":"2025-10-07T20:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.020004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.020055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.020071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.020095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.020112 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.122721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.122768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.122784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.122806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.122823 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.225322 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.225369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.225386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.225413 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.225429 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.328375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.328433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.328450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.328473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.328491 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.431794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.431847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.431863 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.431885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.431903 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.535618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.535663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.535679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.535700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.535716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.638669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.638726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.638743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.638767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.638784 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.678713 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:26 crc kubenswrapper[4926]: E1007 20:57:26.678918 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.741290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.741357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.741376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.741401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.741420 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.844160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.844257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.844274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.844297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.844314 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.947315 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.947374 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.947393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.947416 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:26 crc kubenswrapper[4926]: I1007 20:57:26.947450 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:26Z","lastTransitionTime":"2025-10-07T20:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.050111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.050545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.050666 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.050803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.050922 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.154524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.154576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.154593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.154618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.154635 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.256421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.256749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.256980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.257232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.257505 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.359668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.359714 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.359725 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.359744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.359755 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.462446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.462502 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.462518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.462542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.462572 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.565747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.565787 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.565805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.565826 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.565841 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.668340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.668387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.668404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.668427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.668449 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.677857 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.677894 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.677865 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:27 crc kubenswrapper[4926]: E1007 20:57:27.678015 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:27 crc kubenswrapper[4926]: E1007 20:57:27.678155 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:27 crc kubenswrapper[4926]: E1007 20:57:27.678353 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.772270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.772330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.772346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.772368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.772388 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.875601 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.875665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.875685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.875710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.875728 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.979166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.979254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.979274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.979298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:27 crc kubenswrapper[4926]: I1007 20:57:27.979316 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:27Z","lastTransitionTime":"2025-10-07T20:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.083184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.083349 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.083382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.083417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.083436 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.186049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.186111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.186130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.186154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.186171 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.288573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.288627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.288642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.288667 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.288681 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.391500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.391550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.391559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.391574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.391588 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.494858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.494927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.494946 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.494974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.494991 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.597730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.597776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.597789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.597807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.597820 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.678257 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:28 crc kubenswrapper[4926]: E1007 20:57:28.678930 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.700676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.700746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.700770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.700799 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.700824 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.803547 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.803608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.803630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.803694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.803716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.906405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.906465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.906487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.906514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:28 crc kubenswrapper[4926]: I1007 20:57:28.906535 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:28Z","lastTransitionTime":"2025-10-07T20:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.008987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.009026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.009038 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.009054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.009065 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.111652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.112020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.112239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.112417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.112581 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.215301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.215346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.215363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.215383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.215399 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.265375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.265416 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.265433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.265454 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.265470 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.286007 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.290439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.290473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.290490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.290512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.290530 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.310306 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.314929 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.314978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.314995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.315017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.315033 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.334718 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.339424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.339471 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.339489 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.339512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.339528 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.359361 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.363651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.363732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.363749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.363771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.363789 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.382739 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T20:57:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"86d54cec-6d51-4e27-bc1f-9a644fe10e6e\\\",\\\"systemUUID\\\":\\\"c4ed7d45-a5e6-4ee8-b6f9-5f85ecf4edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T20:57:29Z is after 2025-08-24T17:21:41Z" Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.382965 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.384981 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.385095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.385182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.385304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.385376 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.488389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.488453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.488474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.488501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.488519 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.591378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.591467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.591484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.591507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.591524 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.678606 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.678653 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.678617 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.678829 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.679084 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:29 crc kubenswrapper[4926]: E1007 20:57:29.679191 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.694365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.694401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.694412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.694431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.694443 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.797367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.797423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.797442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.797464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.797481 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.900685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.900744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.900766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.900798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:29 crc kubenswrapper[4926]: I1007 20:57:29.900819 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:29Z","lastTransitionTime":"2025-10-07T20:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.002786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.002854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.002877 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.002905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.002926 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.105509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.105564 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.105584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.105608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.105626 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.208517 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.208585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.208607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.208639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.208656 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.311966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.312020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.312040 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.312066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.312087 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.415857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.415920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.415938 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.415963 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.415980 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.451043 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:30 crc kubenswrapper[4926]: E1007 20:57:30.451246 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:57:30 crc kubenswrapper[4926]: E1007 20:57:30.451328 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs podName:7a8e664c-9b14-4510-8a51-bec46ecbaea5 nodeName:}" failed. No retries permitted until 2025-10-07 20:58:34.451303672 +0000 UTC m=+164.489884832 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs") pod "network-metrics-daemon-c6mr4" (UID: "7a8e664c-9b14-4510-8a51-bec46ecbaea5") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.518675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.518734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.518750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.518775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.518793 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.621286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.621387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.621402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.621422 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.621436 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.678846 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:30 crc kubenswrapper[4926]: E1007 20:57:30.679271 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.724300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.724348 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.724361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.724380 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.724394 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.732646 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=77.732629979 podStartE2EDuration="1m17.732629979s" podCreationTimestamp="2025-10-07 20:56:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.732478515 +0000 UTC m=+100.771059745" watchObservedRunningTime="2025-10-07 20:57:30.732629979 +0000 UTC m=+100.771211139" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.802067 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6dbl" podStartSLOduration=78.802041199 podStartE2EDuration="1m18.802041199s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.801608476 +0000 UTC m=+100.840189636" watchObservedRunningTime="2025-10-07 20:57:30.802041199 +0000 UTC m=+100.840622389" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.802634 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podStartSLOduration=79.802624156 podStartE2EDuration="1m19.802624156s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.775539638 +0000 UTC m=+100.814120838" watchObservedRunningTime="2025-10-07 20:57:30.802624156 +0000 UTC m=+100.841205336" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.828431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.828476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.828492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.828513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.828530 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.851027 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.850996833 podStartE2EDuration="1m20.850996833s" podCreationTimestamp="2025-10-07 20:56:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.828106177 +0000 UTC m=+100.866687357" watchObservedRunningTime="2025-10-07 20:57:30.850996833 +0000 UTC m=+100.889578023" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.873546 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.873519659 podStartE2EDuration="1m18.873519659s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.850785407 +0000 UTC m=+100.889366577" watchObservedRunningTime="2025-10-07 20:57:30.873519659 +0000 UTC m=+100.912100849" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.892915 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.892893923 podStartE2EDuration="47.892893923s" podCreationTimestamp="2025-10-07 20:56:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.874862228 +0000 UTC m=+100.913443398" watchObservedRunningTime="2025-10-07 20:57:30.892893923 +0000 UTC m=+100.931475103" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.931249 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.931468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.931528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.931591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.931645 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:30Z","lastTransitionTime":"2025-10-07T20:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.972712 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-j4btw" podStartSLOduration=79.972693005 podStartE2EDuration="1m19.972693005s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.956310938 +0000 UTC m=+100.994892088" watchObservedRunningTime="2025-10-07 20:57:30.972693005 +0000 UTC m=+101.011274155" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.972827 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2tlv9" podStartSLOduration=78.972823019 podStartE2EDuration="1m18.972823019s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.972676534 +0000 UTC m=+101.011257684" watchObservedRunningTime="2025-10-07 20:57:30.972823019 +0000 UTC m=+101.011404169" Oct 07 20:57:30 crc kubenswrapper[4926]: I1007 20:57:30.993679 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xtr42" podStartSLOduration=78.993653505 podStartE2EDuration="1m18.993653505s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:30.991561214 +0000 UTC m=+101.030142394" watchObservedRunningTime="2025-10-07 20:57:30.993653505 +0000 UTC m=+101.032234675" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.009342 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=35.009312821 podStartE2EDuration="35.009312821s" podCreationTimestamp="2025-10-07 20:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:31.008488607 +0000 UTC m=+101.047069777" watchObservedRunningTime="2025-10-07 20:57:31.009312821 +0000 UTC m=+101.047893991" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.033952 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.034007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.034021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.034044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.034060 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.068468 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-rmrts" podStartSLOduration=80.068446951 podStartE2EDuration="1m20.068446951s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:31.058298596 +0000 UTC m=+101.096879766" watchObservedRunningTime="2025-10-07 20:57:31.068446951 +0000 UTC m=+101.107028101" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.136703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.137080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.137321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.137556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.137828 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.242688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.242736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.242747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.242769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.242781 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.345499 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.345565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.345580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.345968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.346011 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.449376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.449645 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.449743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.449858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.449964 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.553751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.553815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.553838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.553868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.553889 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.658075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.658132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.658158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.658182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.658225 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.678271 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.678341 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.678343 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:31 crc kubenswrapper[4926]: E1007 20:57:31.678455 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:31 crc kubenswrapper[4926]: E1007 20:57:31.678637 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:31 crc kubenswrapper[4926]: E1007 20:57:31.678815 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.761514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.761571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.761581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.761606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.761619 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.864597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.864662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.864680 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.864712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.864732 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.968518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.968594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.968613 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.968641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:31 crc kubenswrapper[4926]: I1007 20:57:31.968661 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:31Z","lastTransitionTime":"2025-10-07T20:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.071677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.072117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.072411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.072498 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.072530 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.175673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.175744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.175762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.175791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.175809 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.278893 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.278968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.278989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.279018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.279038 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.383062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.383598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.383775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.383944 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.384080 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.488182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.488519 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.488542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.488572 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.488594 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.591562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.591622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.591657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.591687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.591706 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.677987 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:32 crc kubenswrapper[4926]: E1007 20:57:32.678410 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.694578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.694652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.694672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.694702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.694722 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.797821 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.798092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.798163 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.798262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.798336 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.900931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.900973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.900983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.900997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:32 crc kubenswrapper[4926]: I1007 20:57:32.901007 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:32Z","lastTransitionTime":"2025-10-07T20:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.003680 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.003735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.003753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.003775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.003792 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.106675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.106723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.106744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.106774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.106796 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.208880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.208925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.208941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.208962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.208980 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.311579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.311874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.312105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.312234 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.312347 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.415872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.416139 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.416223 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.416316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.416378 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.519696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.519761 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.519780 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.519806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.519825 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.623166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.623263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.623289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.623318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.623340 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.678768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.678810 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.678785 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:33 crc kubenswrapper[4926]: E1007 20:57:33.678931 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:33 crc kubenswrapper[4926]: E1007 20:57:33.679053 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:33 crc kubenswrapper[4926]: E1007 20:57:33.679143 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.725941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.725995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.726012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.726034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.726050 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.829578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.829648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.829672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.829701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.829718 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.932475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.932576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.932595 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.932623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:33 crc kubenswrapper[4926]: I1007 20:57:33.932641 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:33Z","lastTransitionTime":"2025-10-07T20:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.036053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.036103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.036119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.036143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.036158 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.138296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.138357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.138378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.138404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.138421 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.241536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.241580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.241590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.241607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.241618 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.345172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.345258 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.345273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.345293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.345306 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.448861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.448926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.448948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.448972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.448990 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.551642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.551687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.551699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.551719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.551732 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.654892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.654954 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.654974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.654998 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.655019 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.678938 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:34 crc kubenswrapper[4926]: E1007 20:57:34.679110 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.758242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.758310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.758327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.758349 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.758366 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.861378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.861441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.861457 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.861479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.861496 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.965298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.965439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.965465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.965492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:34 crc kubenswrapper[4926]: I1007 20:57:34.965514 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:34Z","lastTransitionTime":"2025-10-07T20:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.068812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.068900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.068923 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.068955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.068977 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.172389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.172782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.172986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.173236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.173453 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.276034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.276158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.276172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.276221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.276239 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.379231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.379640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.379816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.379955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.380103 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.483522 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.483590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.483613 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.483642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.483665 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.587646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.587718 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.587739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.587771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.587791 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.678603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:35 crc kubenswrapper[4926]: E1007 20:57:35.678837 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.679142 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:35 crc kubenswrapper[4926]: E1007 20:57:35.679305 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.679533 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:35 crc kubenswrapper[4926]: E1007 20:57:35.679637 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.691677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.691733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.691749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.691768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.691780 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.794244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.794304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.794324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.794351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.794373 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.896654 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.896685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.896693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.896706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.896737 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.999185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.999239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.999251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.999267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:35 crc kubenswrapper[4926]: I1007 20:57:35.999278 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:35Z","lastTransitionTime":"2025-10-07T20:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.102424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.102459 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.102470 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.102482 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.102492 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.205235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.205647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.205812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.205958 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.206093 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.309518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.309564 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.309583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.309607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.309625 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.412073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.412146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.412171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.412238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.412264 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.514183 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.514238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.514254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.514270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.514281 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.616396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.616593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.616804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.616949 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.617018 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.678637 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:36 crc kubenswrapper[4926]: E1007 20:57:36.678893 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.720353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.720397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.720414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.720439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.720457 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.823804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.823857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.823873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.823896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.823915 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.926764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.926825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.926844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.926874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:36 crc kubenswrapper[4926]: I1007 20:57:36.926892 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:36Z","lastTransitionTime":"2025-10-07T20:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.030320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.030382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.030401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.030427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.030445 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.133467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.133552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.133573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.133604 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.133630 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.237013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.237391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.237582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.237767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.238095 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.342567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.342625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.342647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.342675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.342696 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.446016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.446098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.446124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.446156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.446178 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.548989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.549046 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.549063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.549088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.549105 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.667300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.667370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.667391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.667421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.667441 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.677962 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.677972 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.677982 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:37 crc kubenswrapper[4926]: E1007 20:57:37.678113 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:37 crc kubenswrapper[4926]: E1007 20:57:37.678352 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:37 crc kubenswrapper[4926]: E1007 20:57:37.678556 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.770352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.770421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.770439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.770464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.770487 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.873104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.873235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.873261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.873290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.873314 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.976609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.976670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.976687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.976710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:37 crc kubenswrapper[4926]: I1007 20:57:37.976727 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:37Z","lastTransitionTime":"2025-10-07T20:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.079777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.079837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.079857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.079885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.079906 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.183161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.183247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.183259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.183277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.183290 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.285628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.285673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.285687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.285707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.285720 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.388347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.388435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.388477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.388514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.388536 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.491669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.491719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.491736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.491758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.491775 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.595247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.595304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.595322 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.595346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.595363 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.678065 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:38 crc kubenswrapper[4926]: E1007 20:57:38.678538 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.698706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.698800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.698822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.698846 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.698862 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.801464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.801514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.801530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.801551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.801568 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.904297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.904366 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.904389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.904418 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:38 crc kubenswrapper[4926]: I1007 20:57:38.904439 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:38Z","lastTransitionTime":"2025-10-07T20:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.006916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.006988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.007011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.007042 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.007064 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:39Z","lastTransitionTime":"2025-10-07T20:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.110185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.110335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.110359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.110387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.110408 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:39Z","lastTransitionTime":"2025-10-07T20:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.213617 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.213682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.213700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.213724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.213741 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:39Z","lastTransitionTime":"2025-10-07T20:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.316047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.316169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.316240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.316274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.316295 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:39Z","lastTransitionTime":"2025-10-07T20:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.419644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.419723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.419746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.419776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.419798 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:39Z","lastTransitionTime":"2025-10-07T20:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.488594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.488654 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.488674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.488699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.488716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T20:57:39Z","lastTransitionTime":"2025-10-07T20:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.549846 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx"] Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.550926 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.553902 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.554509 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.555001 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.555024 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.654716 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f69740f-c838-4a65-a6ae-2cf320373f05-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.654805 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f69740f-c838-4a65-a6ae-2cf320373f05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.654868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f69740f-c838-4a65-a6ae-2cf320373f05-service-ca\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.654904 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f69740f-c838-4a65-a6ae-2cf320373f05-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.655011 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f69740f-c838-4a65-a6ae-2cf320373f05-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.677961 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:39 crc kubenswrapper[4926]: E1007 20:57:39.678142 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.678411 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.678742 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:39 crc kubenswrapper[4926]: E1007 20:57:39.678948 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:39 crc kubenswrapper[4926]: E1007 20:57:39.679790 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.680041 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 20:57:39 crc kubenswrapper[4926]: E1007 20:57:39.680273 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2zw6c_openshift-ovn-kubernetes(9b8c34cd-2651-4f25-8439-1897f0e937b4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.756010 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f69740f-c838-4a65-a6ae-2cf320373f05-service-ca\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.756061 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f69740f-c838-4a65-a6ae-2cf320373f05-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.756096 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f69740f-c838-4a65-a6ae-2cf320373f05-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.756168 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f69740f-c838-4a65-a6ae-2cf320373f05-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.756261 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f69740f-c838-4a65-a6ae-2cf320373f05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.756346 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f69740f-c838-4a65-a6ae-2cf320373f05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.757163 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f69740f-c838-4a65-a6ae-2cf320373f05-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.757608 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f69740f-c838-4a65-a6ae-2cf320373f05-service-ca\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.765765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f69740f-c838-4a65-a6ae-2cf320373f05-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.789060 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f69740f-c838-4a65-a6ae-2cf320373f05-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-26mwx\" (UID: \"3f69740f-c838-4a65-a6ae-2cf320373f05\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:39 crc kubenswrapper[4926]: I1007 20:57:39.867436 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" Oct 07 20:57:40 crc kubenswrapper[4926]: I1007 20:57:40.299623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" event={"ID":"3f69740f-c838-4a65-a6ae-2cf320373f05","Type":"ContainerStarted","Data":"c1cdfa52d9e9725502279640671bdb90e4432b218dfc2e2274e8331a535769ff"} Oct 07 20:57:40 crc kubenswrapper[4926]: I1007 20:57:40.300077 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" event={"ID":"3f69740f-c838-4a65-a6ae-2cf320373f05","Type":"ContainerStarted","Data":"5188c3f837c8b7bc6fff04bc4c3e6fda686a95d8dcbdc46bbfb591973d7b3872"} Oct 07 20:57:40 crc kubenswrapper[4926]: I1007 20:57:40.321309 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-26mwx" podStartSLOduration=88.321281541 podStartE2EDuration="1m28.321281541s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:40.320541429 +0000 UTC m=+110.359122679" watchObservedRunningTime="2025-10-07 20:57:40.321281541 +0000 UTC m=+110.359862721" Oct 07 20:57:40 crc kubenswrapper[4926]: I1007 20:57:40.678791 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:40 crc kubenswrapper[4926]: E1007 20:57:40.680715 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:41 crc kubenswrapper[4926]: I1007 20:57:41.678342 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:41 crc kubenswrapper[4926]: I1007 20:57:41.678491 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:41 crc kubenswrapper[4926]: I1007 20:57:41.678518 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:41 crc kubenswrapper[4926]: E1007 20:57:41.678615 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:41 crc kubenswrapper[4926]: E1007 20:57:41.678723 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:41 crc kubenswrapper[4926]: E1007 20:57:41.678920 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:42 crc kubenswrapper[4926]: I1007 20:57:42.678698 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:42 crc kubenswrapper[4926]: E1007 20:57:42.680139 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:43 crc kubenswrapper[4926]: I1007 20:57:43.678245 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:43 crc kubenswrapper[4926]: I1007 20:57:43.678281 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:43 crc kubenswrapper[4926]: I1007 20:57:43.678331 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:43 crc kubenswrapper[4926]: E1007 20:57:43.678404 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:43 crc kubenswrapper[4926]: E1007 20:57:43.678551 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:43 crc kubenswrapper[4926]: E1007 20:57:43.678733 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:44 crc kubenswrapper[4926]: I1007 20:57:44.678611 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:44 crc kubenswrapper[4926]: E1007 20:57:44.678812 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:45 crc kubenswrapper[4926]: I1007 20:57:45.678543 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:45 crc kubenswrapper[4926]: I1007 20:57:45.678611 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:45 crc kubenswrapper[4926]: I1007 20:57:45.678546 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:45 crc kubenswrapper[4926]: E1007 20:57:45.678739 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:45 crc kubenswrapper[4926]: E1007 20:57:45.678881 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:45 crc kubenswrapper[4926]: E1007 20:57:45.678947 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.323166 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/1.log" Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.323886 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/0.log" Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.323968 4926 generic.go:334] "Generic (PLEG): container finished" podID="b1217995-75cc-4fc7-8f5f-7752d2e378e2" containerID="31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5" exitCode=1 Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.324011 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerDied","Data":"31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5"} Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.324091 4926 scope.go:117] "RemoveContainer" containerID="8285b8672b0c51645e4116768a8d76b1fdd88ae3709d3f58d79f13b6b5b7e2d6" Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.324535 4926 scope.go:117] "RemoveContainer" containerID="31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5" Oct 07 20:57:46 crc kubenswrapper[4926]: E1007 20:57:46.324781 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2tlv9_openshift-multus(b1217995-75cc-4fc7-8f5f-7752d2e378e2)\"" pod="openshift-multus/multus-2tlv9" podUID="b1217995-75cc-4fc7-8f5f-7752d2e378e2" Oct 07 20:57:46 crc kubenswrapper[4926]: I1007 20:57:46.678068 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:46 crc kubenswrapper[4926]: E1007 20:57:46.678232 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:47 crc kubenswrapper[4926]: I1007 20:57:47.327966 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/1.log" Oct 07 20:57:47 crc kubenswrapper[4926]: I1007 20:57:47.677834 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:47 crc kubenswrapper[4926]: I1007 20:57:47.677857 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:47 crc kubenswrapper[4926]: I1007 20:57:47.678323 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:47 crc kubenswrapper[4926]: E1007 20:57:47.678487 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:47 crc kubenswrapper[4926]: E1007 20:57:47.678740 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:47 crc kubenswrapper[4926]: E1007 20:57:47.678903 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:48 crc kubenswrapper[4926]: I1007 20:57:48.677923 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:48 crc kubenswrapper[4926]: E1007 20:57:48.678101 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:49 crc kubenswrapper[4926]: I1007 20:57:49.678913 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:49 crc kubenswrapper[4926]: I1007 20:57:49.678932 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:49 crc kubenswrapper[4926]: I1007 20:57:49.678932 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:49 crc kubenswrapper[4926]: E1007 20:57:49.679413 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:49 crc kubenswrapper[4926]: E1007 20:57:49.679568 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:49 crc kubenswrapper[4926]: E1007 20:57:49.679089 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:50 crc kubenswrapper[4926]: E1007 20:57:50.663397 4926 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 07 20:57:50 crc kubenswrapper[4926]: I1007 20:57:50.678975 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:50 crc kubenswrapper[4926]: E1007 20:57:50.679895 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:50 crc kubenswrapper[4926]: E1007 20:57:50.777362 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 20:57:51 crc kubenswrapper[4926]: I1007 20:57:51.678426 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:51 crc kubenswrapper[4926]: I1007 20:57:51.678501 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:51 crc kubenswrapper[4926]: I1007 20:57:51.678519 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:51 crc kubenswrapper[4926]: E1007 20:57:51.678611 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:51 crc kubenswrapper[4926]: E1007 20:57:51.678827 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:51 crc kubenswrapper[4926]: E1007 20:57:51.679007 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:52 crc kubenswrapper[4926]: I1007 20:57:52.679837 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 20:57:52 crc kubenswrapper[4926]: I1007 20:57:52.680970 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:52 crc kubenswrapper[4926]: E1007 20:57:52.681273 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.352429 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/3.log" Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.355580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerStarted","Data":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.356157 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.391522 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podStartSLOduration=101.391508208 podStartE2EDuration="1m41.391508208s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:57:53.390455367 +0000 UTC m=+123.429036517" watchObservedRunningTime="2025-10-07 20:57:53.391508208 +0000 UTC m=+123.430089368" Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.463801 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-c6mr4"] Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.463932 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:53 crc kubenswrapper[4926]: E1007 20:57:53.464065 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.678628 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:53 crc kubenswrapper[4926]: I1007 20:57:53.678712 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:53 crc kubenswrapper[4926]: E1007 20:57:53.678746 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:53 crc kubenswrapper[4926]: E1007 20:57:53.678882 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:54 crc kubenswrapper[4926]: I1007 20:57:54.678472 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:54 crc kubenswrapper[4926]: I1007 20:57:54.678555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:54 crc kubenswrapper[4926]: E1007 20:57:54.678996 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:54 crc kubenswrapper[4926]: E1007 20:57:54.679159 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:55 crc kubenswrapper[4926]: I1007 20:57:55.678281 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:55 crc kubenswrapper[4926]: I1007 20:57:55.678277 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:55 crc kubenswrapper[4926]: E1007 20:57:55.678421 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:55 crc kubenswrapper[4926]: E1007 20:57:55.678728 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:55 crc kubenswrapper[4926]: E1007 20:57:55.779412 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 20:57:56 crc kubenswrapper[4926]: I1007 20:57:56.678768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:56 crc kubenswrapper[4926]: I1007 20:57:56.678771 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:56 crc kubenswrapper[4926]: E1007 20:57:56.678946 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:56 crc kubenswrapper[4926]: E1007 20:57:56.679191 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:57 crc kubenswrapper[4926]: I1007 20:57:57.678622 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:57 crc kubenswrapper[4926]: I1007 20:57:57.678636 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:57 crc kubenswrapper[4926]: E1007 20:57:57.678801 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:57 crc kubenswrapper[4926]: E1007 20:57:57.678962 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:58 crc kubenswrapper[4926]: I1007 20:57:58.678850 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:57:58 crc kubenswrapper[4926]: I1007 20:57:58.678910 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:57:58 crc kubenswrapper[4926]: E1007 20:57:58.679004 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:57:58 crc kubenswrapper[4926]: E1007 20:57:58.679086 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:57:59 crc kubenswrapper[4926]: I1007 20:57:59.678029 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:57:59 crc kubenswrapper[4926]: I1007 20:57:59.678131 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:57:59 crc kubenswrapper[4926]: E1007 20:57:59.678694 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:57:59 crc kubenswrapper[4926]: E1007 20:57:59.679055 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:57:59 crc kubenswrapper[4926]: I1007 20:57:59.679608 4926 scope.go:117] "RemoveContainer" containerID="31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5" Oct 07 20:58:00 crc kubenswrapper[4926]: I1007 20:58:00.385770 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/1.log" Oct 07 20:58:00 crc kubenswrapper[4926]: I1007 20:58:00.385872 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerStarted","Data":"466b131acdb2f6ab6515194bccd251a8891942e98beea348991fa0cfa6ba158b"} Oct 07 20:58:00 crc kubenswrapper[4926]: I1007 20:58:00.678702 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:00 crc kubenswrapper[4926]: I1007 20:58:00.678771 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:00 crc kubenswrapper[4926]: E1007 20:58:00.680992 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:58:00 crc kubenswrapper[4926]: E1007 20:58:00.681136 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:58:00 crc kubenswrapper[4926]: E1007 20:58:00.780273 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 20:58:01 crc kubenswrapper[4926]: I1007 20:58:01.678164 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:01 crc kubenswrapper[4926]: I1007 20:58:01.678347 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:01 crc kubenswrapper[4926]: E1007 20:58:01.678441 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:58:01 crc kubenswrapper[4926]: E1007 20:58:01.678650 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:58:02 crc kubenswrapper[4926]: I1007 20:58:02.678225 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:02 crc kubenswrapper[4926]: I1007 20:58:02.678407 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:02 crc kubenswrapper[4926]: E1007 20:58:02.678599 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:58:02 crc kubenswrapper[4926]: E1007 20:58:02.678803 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:58:03 crc kubenswrapper[4926]: I1007 20:58:03.677893 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:03 crc kubenswrapper[4926]: I1007 20:58:03.677904 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:03 crc kubenswrapper[4926]: E1007 20:58:03.678070 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:58:03 crc kubenswrapper[4926]: E1007 20:58:03.678187 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:58:04 crc kubenswrapper[4926]: I1007 20:58:04.678478 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:04 crc kubenswrapper[4926]: I1007 20:58:04.678631 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:04 crc kubenswrapper[4926]: E1007 20:58:04.678674 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 20:58:04 crc kubenswrapper[4926]: E1007 20:58:04.678883 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c6mr4" podUID="7a8e664c-9b14-4510-8a51-bec46ecbaea5" Oct 07 20:58:05 crc kubenswrapper[4926]: I1007 20:58:05.678168 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:05 crc kubenswrapper[4926]: I1007 20:58:05.678289 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:05 crc kubenswrapper[4926]: E1007 20:58:05.678394 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 20:58:05 crc kubenswrapper[4926]: E1007 20:58:05.678469 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 20:58:06 crc kubenswrapper[4926]: I1007 20:58:06.678740 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:06 crc kubenswrapper[4926]: I1007 20:58:06.678803 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:06 crc kubenswrapper[4926]: I1007 20:58:06.681575 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 07 20:58:06 crc kubenswrapper[4926]: I1007 20:58:06.681645 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 07 20:58:06 crc kubenswrapper[4926]: I1007 20:58:06.682165 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 07 20:58:06 crc kubenswrapper[4926]: I1007 20:58:06.682273 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 07 20:58:07 crc kubenswrapper[4926]: I1007 20:58:07.678740 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:07 crc kubenswrapper[4926]: I1007 20:58:07.678867 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:07 crc kubenswrapper[4926]: I1007 20:58:07.681743 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 07 20:58:07 crc kubenswrapper[4926]: I1007 20:58:07.682279 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.174160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.258057 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k87k7"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.258943 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.260919 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.262188 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.262562 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.289400 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.289473 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qzh7t"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.289491 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.290111 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.290388 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.290481 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.290506 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.298216 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.298718 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.299153 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.299691 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.299701 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.300833 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n4dfn"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.301628 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.301761 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.301656 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.302770 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s5gjj"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.303421 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.303629 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.303858 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.304272 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.304478 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.304637 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.304681 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.304947 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.305255 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.305989 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.308274 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ml4zz"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.308923 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.309678 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ggd2q"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.310461 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.310584 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.311099 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.313023 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.313716 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.314390 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6gxjd"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.314763 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.315157 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.315451 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.328289 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-w6j8g"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.328992 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.329560 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.329961 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330458 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330462 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330681 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330468 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330556 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330840 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.330606 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.331281 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-k8kp9"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.332017 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.334979 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335361 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335536 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335644 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335554 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335816 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335743 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335989 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336125 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336307 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336545 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336349 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336571 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336610 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336699 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.337039 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.335365 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.336908 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.337243 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.337420 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.337344 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.337637 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.337833 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.338598 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.338754 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.338912 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339072 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339150 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339263 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339399 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339645 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339784 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.339908 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.340121 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.340277 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.340395 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.341017 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kfjx5"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.341373 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zq6qq"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.342155 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.342527 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.342745 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.361802 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.362376 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k87k7"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.363852 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.363897 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.364075 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.365958 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.366443 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.366695 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.374979 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.375020 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.375327 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.375588 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.375813 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.375990 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.376479 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.376656 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.376820 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.376947 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.376658 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.377211 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7jnfg"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.377801 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.389429 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.389611 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.389752 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.389864 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.389981 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.390079 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.390184 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.390296 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.390424 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.390502 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.391989 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.392159 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.392879 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.395393 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s5gjj"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396383 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396483 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396555 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396626 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396693 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396779 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396890 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.396969 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.397100 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.397567 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.397688 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.397976 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.398493 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.398959 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.403207 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.403495 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.409585 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.409635 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.412026 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qzh7t"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.412058 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5xsvk"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.417063 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.426079 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.427158 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.428432 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.428943 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.431295 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.432158 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.441595 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.441866 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.443090 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.443822 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.444274 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.444543 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.444630 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.444956 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.446321 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.446717 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.447708 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mktfg"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.448155 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.449355 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.449800 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.451734 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7928l"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.452137 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.453313 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-twp9k"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.453719 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.454124 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.455048 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.455610 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.456294 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.456607 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.459987 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ml4zz"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.460027 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.460417 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.460563 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.460679 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.461221 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.461874 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.462672 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.463957 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w6j8g"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465168 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n4dfn"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465464 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-ca\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465487 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aaea817a-b104-4b5f-a074-e7196d3ec448-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d353344-979a-4bb8-9c0c-04c073ca02db-serving-cert\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465524 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqzwh\" (UniqueName: \"kubernetes.io/projected/4d353344-979a-4bb8-9c0c-04c073ca02db-kube-api-access-jqzwh\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465540 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-config\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465557 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64bdf46c-1845-4807-8dfa-4c9526181afc-config\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465574 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6r72\" (UniqueName: \"kubernetes.io/projected/64bdf46c-1845-4807-8dfa-4c9526181afc-kube-api-access-m6r72\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465743 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64bdf46c-1845-4807-8dfa-4c9526181afc-serving-cert\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465768 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-client\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465792 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87a2e5d7-ad73-4643-b301-5424d27d6e55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465810 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-service-ca\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaea817a-b104-4b5f-a074-e7196d3ec448-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465874 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvfhl\" (UniqueName: \"kubernetes.io/projected/bf45cac1-ea89-474f-8596-e895d22be327-kube-api-access-kvfhl\") pod \"downloads-7954f5f757-w6j8g\" (UID: \"bf45cac1-ea89-474f-8596-e895d22be327\") " pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87a2e5d7-ad73-4643-b301-5424d27d6e55-serving-cert\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssfkn\" (UniqueName: \"kubernetes.io/projected/aaea817a-b104-4b5f-a074-e7196d3ec448-kube-api-access-ssfkn\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465922 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/64bdf46c-1845-4807-8dfa-4c9526181afc-trusted-ca\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.465962 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mhtd\" (UniqueName: \"kubernetes.io/projected/87a2e5d7-ad73-4643-b301-5424d27d6e55-kube-api-access-6mhtd\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.466844 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-k8kp9"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.468132 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z7z5t"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.469339 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.471590 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zq6qq"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.473117 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.474931 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.476937 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.478031 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.479480 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.481080 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5xsvk"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.482062 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7jnfg"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.482140 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.482962 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.483991 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6gxjd"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.484955 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.487020 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.487047 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kfjx5"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.488582 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.489274 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ggd2q"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.490041 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.491263 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.491861 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.492811 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mktfg"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.493783 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.494754 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.495909 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.497271 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.505178 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.508257 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-xg824"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.509361 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qdvtq"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.509533 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.509945 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z7z5t"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.510006 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.510859 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.511666 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-twp9k"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.512660 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xg824"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.513809 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.515086 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qdvtq"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.516111 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8cwxz"] Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.516621 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.521225 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.541260 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.560514 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567115 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87a2e5d7-ad73-4643-b301-5424d27d6e55-serving-cert\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567153 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssfkn\" (UniqueName: \"kubernetes.io/projected/aaea817a-b104-4b5f-a074-e7196d3ec448-kube-api-access-ssfkn\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/64bdf46c-1845-4807-8dfa-4c9526181afc-trusted-ca\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567250 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mhtd\" (UniqueName: \"kubernetes.io/projected/87a2e5d7-ad73-4643-b301-5424d27d6e55-kube-api-access-6mhtd\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567275 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-ca\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567297 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aaea817a-b104-4b5f-a074-e7196d3ec448-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567321 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d353344-979a-4bb8-9c0c-04c073ca02db-serving-cert\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567343 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqzwh\" (UniqueName: \"kubernetes.io/projected/4d353344-979a-4bb8-9c0c-04c073ca02db-kube-api-access-jqzwh\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567366 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-config\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567389 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64bdf46c-1845-4807-8dfa-4c9526181afc-config\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567411 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6r72\" (UniqueName: \"kubernetes.io/projected/64bdf46c-1845-4807-8dfa-4c9526181afc-kube-api-access-m6r72\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567437 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64bdf46c-1845-4807-8dfa-4c9526181afc-serving-cert\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567456 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-client\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567481 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87a2e5d7-ad73-4643-b301-5424d27d6e55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567504 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-service-ca\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567542 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaea817a-b104-4b5f-a074-e7196d3ec448-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.567575 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvfhl\" (UniqueName: \"kubernetes.io/projected/bf45cac1-ea89-474f-8596-e895d22be327-kube-api-access-kvfhl\") pod \"downloads-7954f5f757-w6j8g\" (UID: \"bf45cac1-ea89-474f-8596-e895d22be327\") " pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.570559 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87a2e5d7-ad73-4643-b301-5424d27d6e55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.571019 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/64bdf46c-1845-4807-8dfa-4c9526181afc-trusted-ca\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.571514 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaea817a-b104-4b5f-a074-e7196d3ec448-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.571751 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64bdf46c-1845-4807-8dfa-4c9526181afc-config\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.572436 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aaea817a-b104-4b5f-a074-e7196d3ec448-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.572628 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d353344-979a-4bb8-9c0c-04c073ca02db-serving-cert\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.573681 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64bdf46c-1845-4807-8dfa-4c9526181afc-serving-cert\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.581753 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.585860 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87a2e5d7-ad73-4643-b301-5424d27d6e55-serving-cert\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.590349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-client\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.600798 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.608606 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-config\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.621290 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.628416 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-ca\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.640497 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.648635 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d353344-979a-4bb8-9c0c-04c073ca02db-etcd-service-ca\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.661372 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.680815 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.700985 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.727004 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.741164 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.761264 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.780694 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.801730 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.820717 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.840961 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.861330 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.901657 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.921765 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.982863 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 07 20:58:10 crc kubenswrapper[4926]: I1007 20:58:10.983773 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.001913 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.021752 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.042273 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.061319 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.081961 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.102223 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.121553 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.141832 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.161808 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.182041 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.201232 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.221874 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.242872 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.261731 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.281028 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.301915 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.321377 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.341250 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.361106 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.381749 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.460960 4926 request.go:700] Waited for 1.012478723s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.463744 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.463767 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.463966 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.464576 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.481057 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.501721 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.521730 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.542172 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.567651 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.582383 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.601597 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.621424 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.641404 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.662064 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.680998 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.701005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.722022 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.742018 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.762474 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.782045 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.801446 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.821860 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.842033 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.860902 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.882512 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.902615 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.921895 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.942271 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.961635 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 07 20:58:11 crc kubenswrapper[4926]: I1007 20:58:11.982167 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.004610 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.022163 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.042888 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.061274 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.083752 4926 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.102323 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.122772 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.143028 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.161968 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.182697 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.202492 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.221853 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.242918 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.261811 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.283873 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.306616 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.360296 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssfkn\" (UniqueName: \"kubernetes.io/projected/aaea817a-b104-4b5f-a074-e7196d3ec448-kube-api-access-ssfkn\") pod \"openshift-controller-manager-operator-756b6f6bc6-9l6nb\" (UID: \"aaea817a-b104-4b5f-a074-e7196d3ec448\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.379227 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mhtd\" (UniqueName: \"kubernetes.io/projected/87a2e5d7-ad73-4643-b301-5424d27d6e55-kube-api-access-6mhtd\") pod \"openshift-config-operator-7777fb866f-mq6kz\" (UID: \"87a2e5d7-ad73-4643-b301-5424d27d6e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.390647 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqzwh\" (UniqueName: \"kubernetes.io/projected/4d353344-979a-4bb8-9c0c-04c073ca02db-kube-api-access-jqzwh\") pod \"etcd-operator-b45778765-zq6qq\" (UID: \"4d353344-979a-4bb8-9c0c-04c073ca02db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.416224 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvfhl\" (UniqueName: \"kubernetes.io/projected/bf45cac1-ea89-474f-8596-e895d22be327-kube-api-access-kvfhl\") pod \"downloads-7954f5f757-w6j8g\" (UID: \"bf45cac1-ea89-474f-8596-e895d22be327\") " pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.428833 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6r72\" (UniqueName: \"kubernetes.io/projected/64bdf46c-1845-4807-8dfa-4c9526181afc-kube-api-access-m6r72\") pod \"console-operator-58897d9998-6gxjd\" (UID: \"64bdf46c-1845-4807-8dfa-4c9526181afc\") " pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487670 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-service-ca-bundle\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487709 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs94d\" (UniqueName: \"kubernetes.io/projected/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-kube-api-access-xs94d\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-policies\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487791 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/24f2e598-e97c-4583-bc67-cd8afcf6065e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487808 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8c79c29-7251-4b50-8bb1-c066cd8e3060-config\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487822 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2830f943-9ae7-428e-8194-35e921cf4d8f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ec7685a-e96e-434f-be3a-cda923c04ba1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487854 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-config\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-config\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.487970 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-trusted-ca-bundle\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488017 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-oauth-serving-cert\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488120 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-audit\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488222 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488278 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/492b14c9-39b0-4073-ac89-7dad77ecb8be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-465q2\" (UID: \"492b14c9-39b0-4073-ac89-7dad77ecb8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488379 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8c79c29-7251-4b50-8bb1-c066cd8e3060-machine-approver-tls\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488428 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0633e0a1-8695-4237-99fc-a7576cc66d15-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488471 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488517 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37fc498-33f9-4968-8c6a-1cfb00e123b8-serving-cert\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkjvr\" (UniqueName: \"kubernetes.io/projected/b37fc498-33f9-4968-8c6a-1cfb00e123b8-kube-api-access-hkjvr\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488685 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcdxj\" (UniqueName: \"kubernetes.io/projected/24f2e598-e97c-4583-bc67-cd8afcf6065e-kube-api-access-zcdxj\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488725 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-serving-cert\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488802 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-trusted-ca\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21029c24-0c5d-4186-8941-9db6c2d0e31b-audit-dir\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488846 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-image-import-ca\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488865 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-config\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488890 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488952 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qvcn\" (UniqueName: \"kubernetes.io/projected/6d6919d5-336c-452e-94dd-f37223a44f2c-kube-api-access-6qvcn\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488973 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-service-ca\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.488995 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbqv8\" (UniqueName: \"kubernetes.io/projected/f8c79c29-7251-4b50-8bb1-c066cd8e3060-kube-api-access-fbqv8\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-client-ca\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489055 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-bound-sa-token\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489074 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8c79c29-7251-4b50-8bb1-c066cd8e3060-auth-proxy-config\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489128 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489161 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ec7685a-e96e-434f-be3a-cda923c04ba1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489211 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-etcd-client\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489244 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.489252 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:12.989233545 +0000 UTC m=+143.027814825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489278 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9rhp\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-kube-api-access-p9rhp\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-etcd-serving-ca\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489325 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-config\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489412 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-client-ca\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489433 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hpv5\" (UniqueName: \"kubernetes.io/projected/2830f943-9ae7-428e-8194-35e921cf4d8f-kube-api-access-5hpv5\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489454 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24f2e598-e97c-4583-bc67-cd8afcf6065e-config\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489475 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0633e0a1-8695-4237-99fc-a7576cc66d15-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489495 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489520 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-684cm\" (UniqueName: \"kubernetes.io/projected/21029c24-0c5d-4186-8941-9db6c2d0e31b-kube-api-access-684cm\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489540 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489586 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-etcd-client\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489605 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-audit-policies\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489624 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28ccl\" (UniqueName: \"kubernetes.io/projected/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-kube-api-access-28ccl\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfvlm\" (UniqueName: \"kubernetes.io/projected/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-kube-api-access-mfvlm\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4eff0e24-4bcf-44be-85a0-15b8311e39bf-metrics-tls\") pod \"dns-operator-744455d44c-k8kp9\" (UID: \"4eff0e24-4bcf-44be-85a0-15b8311e39bf\") " pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489738 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489758 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df5bv\" (UniqueName: \"kubernetes.io/projected/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-kube-api-access-df5bv\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489776 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2830f943-9ae7-428e-8194-35e921cf4d8f-trusted-ca\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489820 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d6919d5-336c-452e-94dd-f37223a44f2c-serving-cert\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-tls\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489872 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-serving-cert\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489913 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-config\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489938 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-encryption-config\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489969 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-serving-cert\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.489990 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbl85\" (UniqueName: \"kubernetes.io/projected/2f2abaf1-1c6b-4768-8228-2cdbf730a457-kube-api-access-wbl85\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490014 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-dir\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490036 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490056 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-oauth-config\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490074 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-certificates\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490093 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/0633e0a1-8695-4237-99fc-a7576cc66d15-kube-api-access-h8kw9\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490113 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490153 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-serving-cert\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490176 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490217 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0633e0a1-8695-4237-99fc-a7576cc66d15-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490244 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f2abaf1-1c6b-4768-8228-2cdbf730a457-audit-dir\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2830f943-9ae7-428e-8194-35e921cf4d8f-metrics-tls\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490287 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98t8z\" (UniqueName: \"kubernetes.io/projected/492b14c9-39b0-4073-ac89-7dad77ecb8be-kube-api-access-98t8z\") pod \"cluster-samples-operator-665b6dd947-465q2\" (UID: \"492b14c9-39b0-4073-ac89-7dad77ecb8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490308 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f2abaf1-1c6b-4768-8228-2cdbf730a457-node-pullsecrets\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490348 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490368 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-encryption-config\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490385 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/24f2e598-e97c-4583-bc67-cd8afcf6065e-images\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.490416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nchb\" (UniqueName: \"kubernetes.io/projected/4eff0e24-4bcf-44be-85a0-15b8311e39bf-kube-api-access-2nchb\") pod \"dns-operator-744455d44c-k8kp9\" (UID: \"4eff0e24-4bcf-44be-85a0-15b8311e39bf\") " pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.583012 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.589619 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.590853 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591130 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbl85\" (UniqueName: \"kubernetes.io/projected/2f2abaf1-1c6b-4768-8228-2cdbf730a457-kube-api-access-wbl85\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591191 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-dir\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591273 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-oauth-config\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591366 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-certificates\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591422 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591463 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-serving-cert\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591518 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591575 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09517b43-f8f3-4a06-8dbf-0e24e3acc173-tmpfs\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591622 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dc6d7b79-a9de-4462-97a3-2bf2de30f825-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f2abaf1-1c6b-4768-8228-2cdbf730a457-node-pullsecrets\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591717 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98t8z\" (UniqueName: \"kubernetes.io/projected/492b14c9-39b0-4073-ac89-7dad77ecb8be-kube-api-access-98t8z\") pod \"cluster-samples-operator-665b6dd947-465q2\" (UID: \"492b14c9-39b0-4073-ac89-7dad77ecb8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591767 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61924da4-8b4e-4ea9-920b-97252db4a705-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591907 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nchb\" (UniqueName: \"kubernetes.io/projected/4eff0e24-4bcf-44be-85a0-15b8311e39bf-kube-api-access-2nchb\") pod \"dns-operator-744455d44c-k8kp9\" (UID: \"4eff0e24-4bcf-44be-85a0-15b8311e39bf\") " pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.591952 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-registration-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592000 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-service-ca-bundle\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592048 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs94d\" (UniqueName: \"kubernetes.io/projected/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-kube-api-access-xs94d\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-policies\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/24f2e598-e97c-4583-bc67-cd8afcf6065e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwc7l\" (UniqueName: \"kubernetes.io/projected/5026deaa-f35e-4c06-9c90-cff95d6599a2-kube-api-access-nwc7l\") pod \"control-plane-machine-set-operator-78cbb6b69f-6mlnn\" (UID: \"5026deaa-f35e-4c06-9c90-cff95d6599a2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592308 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ec7685a-e96e-434f-be3a-cda923c04ba1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592358 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-csi-data-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592410 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592463 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592516 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43e33278-4035-4e7a-936c-481775481acf-profile-collector-cert\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592569 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d222a2b1-b28d-47d4-98bd-cb7b0456c724-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592654 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0633e0a1-8695-4237-99fc-a7576cc66d15-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/492b14c9-39b0-4073-ac89-7dad77ecb8be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-465q2\" (UID: \"492b14c9-39b0-4073-ac89-7dad77ecb8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592769 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592823 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592885 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37fc498-33f9-4968-8c6a-1cfb00e123b8-serving-cert\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.592954 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkjvr\" (UniqueName: \"kubernetes.io/projected/b37fc498-33f9-4968-8c6a-1cfb00e123b8-kube-api-access-hkjvr\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593000 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09517b43-f8f3-4a06-8dbf-0e24e3acc173-apiservice-cert\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593073 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9qcr\" (UniqueName: \"kubernetes.io/projected/929c455e-fb87-4885-8612-6888167af1d4-kube-api-access-s9qcr\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593127 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21029c24-0c5d-4186-8941-9db6c2d0e31b-audit-dir\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593179 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-image-import-ca\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5026deaa-f35e-4c06-9c90-cff95d6599a2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6mlnn\" (UID: \"5026deaa-f35e-4c06-9c90-cff95d6599a2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593393 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-service-ca\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593440 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593497 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbqv8\" (UniqueName: \"kubernetes.io/projected/f8c79c29-7251-4b50-8bb1-c066cd8e3060-kube-api-access-fbqv8\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43e33278-4035-4e7a-936c-481775481acf-srv-cert\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d222a2b1-b28d-47d4-98bd-cb7b0456c724-config\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593655 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7d810b4-fe88-422e-b3a8-cc2c407039f1-signing-key\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593775 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593838 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd2rw\" (UniqueName: \"kubernetes.io/projected/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-kube-api-access-vd2rw\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593853 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f2abaf1-1c6b-4768-8228-2cdbf730a457-node-pullsecrets\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.593930 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.594026 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.594121 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-dir\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.594145 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.594373 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.094338819 +0000 UTC m=+143.132920149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.594857 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-certificates\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.595996 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.596949 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-policies\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.596966 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.596977 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21029c24-0c5d-4186-8941-9db6c2d0e31b-audit-dir\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.598689 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.598753 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-etcd-serving-ca\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.598836 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.598897 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hpv5\" (UniqueName: \"kubernetes.io/projected/2830f943-9ae7-428e-8194-35e921cf4d8f-kube-api-access-5hpv5\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.598960 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24f2e598-e97c-4583-bc67-cd8afcf6065e-config\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599023 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-socket-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599084 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/eeebe27e-48fb-47b5-bcf4-a640164e4f1e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-57mj8\" (UID: \"eeebe27e-48fb-47b5-bcf4-a640164e4f1e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lgt9\" (UniqueName: \"kubernetes.io/projected/738dcd07-f77b-4737-8195-07e674320384-kube-api-access-6lgt9\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599176 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599276 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-plugins-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599353 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599443 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-mountpoint-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599504 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-etcd-client\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4eff0e24-4bcf-44be-85a0-15b8311e39bf-metrics-tls\") pod \"dns-operator-744455d44c-k8kp9\" (UID: \"4eff0e24-4bcf-44be-85a0-15b8311e39bf\") " pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599877 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-service-ca-bundle\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599892 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-audit-policies\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.599986 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28ccl\" (UniqueName: \"kubernetes.io/projected/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-kube-api-access-28ccl\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600039 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfvlm\" (UniqueName: \"kubernetes.io/projected/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-kube-api-access-mfvlm\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600095 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbrzb\" (UniqueName: \"kubernetes.io/projected/5843662e-8938-4478-a6ab-f92c4a2e75d4-kube-api-access-mbrzb\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600120 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-service-ca\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600158 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df5bv\" (UniqueName: \"kubernetes.io/projected/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-kube-api-access-df5bv\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600243 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2830f943-9ae7-428e-8194-35e921cf4d8f-trusted-ca\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.596265 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600294 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/738dcd07-f77b-4737-8195-07e674320384-secret-volume\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600348 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-serving-cert\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600396 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-encryption-config\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600432 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr8hl\" (UniqueName: \"kubernetes.io/projected/18d2ab0f-cb2d-41ed-b746-28adc1888942-kube-api-access-kr8hl\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600742 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-serving-cert\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.600915 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-image-import-ca\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601450 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-serving-cert\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601483 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-audit-policies\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601494 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-stats-auth\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601594 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601674 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-config-volume\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601534 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24f2e598-e97c-4583-bc67-cd8afcf6065e-config\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.601927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.602032 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ec7685a-e96e-434f-be3a-cda923c04ba1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.602495 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.602575 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/0633e0a1-8695-4237-99fc-a7576cc66d15-kube-api-access-h8kw9\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.602631 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0633e0a1-8695-4237-99fc-a7576cc66d15-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.602735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f2abaf1-1c6b-4768-8228-2cdbf730a457-audit-dir\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.603141 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2830f943-9ae7-428e-8194-35e921cf4d8f-trusted-ca\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.603316 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.603429 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-oauth-config\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.603545 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f2abaf1-1c6b-4768-8228-2cdbf730a457-audit-dir\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604079 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2830f943-9ae7-428e-8194-35e921cf4d8f-metrics-tls\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604424 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0633e0a1-8695-4237-99fc-a7576cc66d15-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604429 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e9895f1f-054d-4b3c-b55d-a02514f51a78-cert\") pod \"ingress-canary-xg824\" (UID: \"e9895f1f-054d-4b3c-b55d-a02514f51a78\") " pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sxft\" (UniqueName: \"kubernetes.io/projected/a7d810b4-fe88-422e-b3a8-cc2c407039f1-kube-api-access-7sxft\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604908 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21029c24-0c5d-4186-8941-9db6c2d0e31b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.604977 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7d810b4-fe88-422e-b3a8-cc2c407039f1-signing-cabundle\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.605031 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-encryption-config\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.605081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/24f2e598-e97c-4583-bc67-cd8afcf6065e-images\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.605127 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rd84\" (UniqueName: \"kubernetes.io/projected/43e33278-4035-4e7a-936c-481775481acf-kube-api-access-9rd84\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.605152 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-images\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.605201 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-default-certificate\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.605586 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-etcd-serving-ca\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.606018 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37fc498-33f9-4968-8c6a-1cfb00e123b8-serving-cert\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.606150 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/24f2e598-e97c-4583-bc67-cd8afcf6065e-images\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.606690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-serving-cert\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607063 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/24f2e598-e97c-4583-bc67-cd8afcf6065e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607251 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-config\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-encryption-config\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607331 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-config\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607424 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8c79c29-7251-4b50-8bb1-c066cd8e3060-config\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607601 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2830f943-9ae7-428e-8194-35e921cf4d8f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607656 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d2ab0f-cb2d-41ed-b746-28adc1888942-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607694 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/44ad9726-b93f-4dd3-bbdf-b4f6945179a3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7jnfg\" (UID: \"44ad9726-b93f-4dd3-bbdf-b4f6945179a3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607793 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-audit\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607877 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-trusted-ca-bundle\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.607919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-oauth-serving-cert\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608048 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpnk6\" (UniqueName: \"kubernetes.io/projected/dc6d7b79-a9de-4462-97a3-2bf2de30f825-kube-api-access-xpnk6\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608094 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5843662e-8938-4478-a6ab-f92c4a2e75d4-serving-cert\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608140 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8c79c29-7251-4b50-8bb1-c066cd8e3060-machine-approver-tls\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d222a2b1-b28d-47d4-98bd-cb7b0456c724-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608262 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mwqc\" (UniqueName: \"kubernetes.io/projected/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-kube-api-access-6mwqc\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-config\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608400 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-metrics-certs\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608446 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59p9r\" (UniqueName: \"kubernetes.io/projected/1680115c-0bb3-4696-8564-a685fe8efdcb-kube-api-access-59p9r\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608481 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rctbf\" (UniqueName: \"kubernetes.io/projected/44ad9726-b93f-4dd3-bbdf-b4f6945179a3-kube-api-access-rctbf\") pod \"multus-admission-controller-857f4d67dd-7jnfg\" (UID: \"44ad9726-b93f-4dd3-bbdf-b4f6945179a3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608528 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcdxj\" (UniqueName: \"kubernetes.io/projected/24f2e598-e97c-4583-bc67-cd8afcf6065e-kube-api-access-zcdxj\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608547 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-config\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608575 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dc6d7b79-a9de-4462-97a3-2bf2de30f825-proxy-tls\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608609 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-metrics-tls\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-trusted-ca\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2f2abaf1-1c6b-4768-8228-2cdbf730a457-audit\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.608686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-serving-cert\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.609695 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-trusted-ca\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-config\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610095 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610305 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qvcn\" (UniqueName: \"kubernetes.io/projected/6d6919d5-336c-452e-94dd-f37223a44f2c-kube-api-access-6qvcn\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610364 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-bound-sa-token\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610369 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/21029c24-0c5d-4186-8941-9db6c2d0e31b-etcd-client\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610401 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8c79c29-7251-4b50-8bb1-c066cd8e3060-auth-proxy-config\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610473 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-oauth-serving-cert\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-client-ca\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610627 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplwq\" (UniqueName: \"kubernetes.io/projected/09517b43-f8f3-4a06-8dbf-0e24e3acc173-kube-api-access-gplwq\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610662 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-srv-cert\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610697 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610745 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610779 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5843662e-8938-4478-a6ab-f92c4a2e75d4-config\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610815 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ec7685a-e96e-434f-be3a-cda923c04ba1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-etcd-client\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.610890 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61924da4-8b4e-4ea9-920b-97252db4a705-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611308 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4eff0e24-4bcf-44be-85a0-15b8311e39bf-metrics-tls\") pod \"dns-operator-744455d44c-k8kp9\" (UID: \"4eff0e24-4bcf-44be-85a0-15b8311e39bf\") " pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611632 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg96q\" (UniqueName: \"kubernetes.io/projected/bd102fa8-8a39-4a85-ac8b-6f16720e4023-kube-api-access-tg96q\") pod \"migrator-59844c95c7-549d6\" (UID: \"bd102fa8-8a39-4a85-ac8b-6f16720e4023\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/492b14c9-39b0-4073-ac89-7dad77ecb8be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-465q2\" (UID: \"492b14c9-39b0-4073-ac89-7dad77ecb8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611702 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9vst\" (UniqueName: \"kubernetes.io/projected/e9895f1f-054d-4b3c-b55d-a02514f51a78-kube-api-access-j9vst\") pod \"ingress-canary-xg824\" (UID: \"e9895f1f-054d-4b3c-b55d-a02514f51a78\") " pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611762 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbbwh\" (UniqueName: \"kubernetes.io/projected/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-kube-api-access-rbbwh\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9rhp\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-kube-api-access-p9rhp\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612139 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsgjq\" (UniqueName: \"kubernetes.io/projected/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-kube-api-access-vsgjq\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612390 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-client-ca\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-config\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612618 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-proxy-tls\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0633e0a1-8695-4237-99fc-a7576cc66d15-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612870 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-serving-cert\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.611852 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8c79c29-7251-4b50-8bb1-c066cd8e3060-auth-proxy-config\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.612995 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613049 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-684cm\" (UniqueName: \"kubernetes.io/projected/21029c24-0c5d-4186-8941-9db6c2d0e31b-kube-api-access-684cm\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613075 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1680115c-0bb3-4696-8564-a685fe8efdcb-service-ca-bundle\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613111 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/738dcd07-f77b-4737-8195-07e674320384-config-volume\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613161 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613214 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-node-bootstrap-token\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613239 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61924da4-8b4e-4ea9-920b-97252db4a705-config\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613289 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d2ab0f-cb2d-41ed-b746-28adc1888942-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613355 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d6919d5-336c-452e-94dd-f37223a44f2c-serving-cert\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613386 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613410 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-tls\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613438 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09517b43-f8f3-4a06-8dbf-0e24e3acc173-webhook-cert\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613470 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc9dj\" (UniqueName: \"kubernetes.io/projected/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-kube-api-access-cc9dj\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613495 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-config\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613523 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-certs\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.613706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d46d\" (UniqueName: \"kubernetes.io/projected/eeebe27e-48fb-47b5-bcf4-a640164e4f1e-kube-api-access-7d46d\") pod \"package-server-manager-789f6589d5-57mj8\" (UID: \"eeebe27e-48fb-47b5-bcf4-a640164e4f1e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.614394 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-config\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.614991 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.114965772 +0000 UTC m=+143.153546932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.615612 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8c79c29-7251-4b50-8bb1-c066cd8e3060-config\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.615895 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2830f943-9ae7-428e-8194-35e921cf4d8f-metrics-tls\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.616137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.616160 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0633e0a1-8695-4237-99fc-a7576cc66d15-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.617759 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-client-ca\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.618387 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-client-ca\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.618698 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.620017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-config\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.620810 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.621536 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-trusted-ca-bundle\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.623767 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ec7685a-e96e-434f-be3a-cda923c04ba1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.624128 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d6919d5-336c-452e-94dd-f37223a44f2c-serving-cert\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.624837 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-config\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.625856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-etcd-client\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.626340 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f2abaf1-1c6b-4768-8228-2cdbf730a457-encryption-config\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.626444 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-serving-cert\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.626485 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8c79c29-7251-4b50-8bb1-c066cd8e3060-machine-approver-tls\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.626444 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-tls\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.627351 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.627423 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.627544 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.628763 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.630468 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.637576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbl85\" (UniqueName: \"kubernetes.io/projected/2f2abaf1-1c6b-4768-8228-2cdbf730a457-kube-api-access-wbl85\") pod \"apiserver-76f77b778f-n4dfn\" (UID: \"2f2abaf1-1c6b-4768-8228-2cdbf730a457\") " pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.648266 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.668353 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98t8z\" (UniqueName: \"kubernetes.io/projected/492b14c9-39b0-4073-ac89-7dad77ecb8be-kube-api-access-98t8z\") pod \"cluster-samples-operator-665b6dd947-465q2\" (UID: \"492b14c9-39b0-4073-ac89-7dad77ecb8be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.685749 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkjvr\" (UniqueName: \"kubernetes.io/projected/b37fc498-33f9-4968-8c6a-1cfb00e123b8-kube-api-access-hkjvr\") pod \"controller-manager-879f6c89f-k87k7\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.700770 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfvlm\" (UniqueName: \"kubernetes.io/projected/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-kube-api-access-mfvlm\") pod \"console-f9d7485db-ggd2q\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716345 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.716553 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.216519088 +0000 UTC m=+143.255100238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716626 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplwq\" (UniqueName: \"kubernetes.io/projected/09517b43-f8f3-4a06-8dbf-0e24e3acc173-kube-api-access-gplwq\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716661 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-srv-cert\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716682 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716717 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716770 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5843662e-8938-4478-a6ab-f92c4a2e75d4-config\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716793 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61924da4-8b4e-4ea9-920b-97252db4a705-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716818 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg96q\" (UniqueName: \"kubernetes.io/projected/bd102fa8-8a39-4a85-ac8b-6f16720e4023-kube-api-access-tg96q\") pod \"migrator-59844c95c7-549d6\" (UID: \"bd102fa8-8a39-4a85-ac8b-6f16720e4023\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9vst\" (UniqueName: \"kubernetes.io/projected/e9895f1f-054d-4b3c-b55d-a02514f51a78-kube-api-access-j9vst\") pod \"ingress-canary-xg824\" (UID: \"e9895f1f-054d-4b3c-b55d-a02514f51a78\") " pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716863 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbbwh\" (UniqueName: \"kubernetes.io/projected/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-kube-api-access-rbbwh\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716895 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsgjq\" (UniqueName: \"kubernetes.io/projected/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-kube-api-access-vsgjq\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716914 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-proxy-tls\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716934 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716967 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1680115c-0bb3-4696-8564-a685fe8efdcb-service-ca-bundle\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.716990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/738dcd07-f77b-4737-8195-07e674320384-config-volume\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717014 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-node-bootstrap-token\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717039 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61924da4-8b4e-4ea9-920b-97252db4a705-config\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717058 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d2ab0f-cb2d-41ed-b746-28adc1888942-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09517b43-f8f3-4a06-8dbf-0e24e3acc173-webhook-cert\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc9dj\" (UniqueName: \"kubernetes.io/projected/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-kube-api-access-cc9dj\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717109 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-certs\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717151 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d46d\" (UniqueName: \"kubernetes.io/projected/eeebe27e-48fb-47b5-bcf4-a640164e4f1e-kube-api-access-7d46d\") pod \"package-server-manager-789f6589d5-57mj8\" (UID: \"eeebe27e-48fb-47b5-bcf4-a640164e4f1e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717169 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09517b43-f8f3-4a06-8dbf-0e24e3acc173-tmpfs\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717224 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dc6d7b79-a9de-4462-97a3-2bf2de30f825-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717241 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61924da4-8b4e-4ea9-920b-97252db4a705-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717280 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-registration-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717309 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwc7l\" (UniqueName: \"kubernetes.io/projected/5026deaa-f35e-4c06-9c90-cff95d6599a2-kube-api-access-nwc7l\") pod \"control-plane-machine-set-operator-78cbb6b69f-6mlnn\" (UID: \"5026deaa-f35e-4c06-9c90-cff95d6599a2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717328 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-csi-data-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717345 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43e33278-4035-4e7a-936c-481775481acf-profile-collector-cert\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717366 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d222a2b1-b28d-47d4-98bd-cb7b0456c724-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717415 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09517b43-f8f3-4a06-8dbf-0e24e3acc173-apiservice-cert\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717448 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9qcr\" (UniqueName: \"kubernetes.io/projected/929c455e-fb87-4885-8612-6888167af1d4-kube-api-access-s9qcr\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717474 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5026deaa-f35e-4c06-9c90-cff95d6599a2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6mlnn\" (UID: \"5026deaa-f35e-4c06-9c90-cff95d6599a2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717508 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43e33278-4035-4e7a-936c-481775481acf-srv-cert\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717530 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d222a2b1-b28d-47d4-98bd-cb7b0456c724-config\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717551 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717570 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7d810b4-fe88-422e-b3a8-cc2c407039f1-signing-key\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717591 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd2rw\" (UniqueName: \"kubernetes.io/projected/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-kube-api-access-vd2rw\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717614 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717636 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717659 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/eeebe27e-48fb-47b5-bcf4-a640164e4f1e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-57mj8\" (UID: \"eeebe27e-48fb-47b5-bcf4-a640164e4f1e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717681 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lgt9\" (UniqueName: \"kubernetes.io/projected/738dcd07-f77b-4737-8195-07e674320384-kube-api-access-6lgt9\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717712 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-socket-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717754 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-plugins-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-mountpoint-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717791 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbrzb\" (UniqueName: \"kubernetes.io/projected/5843662e-8938-4478-a6ab-f92c4a2e75d4-kube-api-access-mbrzb\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717832 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/738dcd07-f77b-4737-8195-07e674320384-secret-volume\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr8hl\" (UniqueName: \"kubernetes.io/projected/18d2ab0f-cb2d-41ed-b746-28adc1888942-kube-api-access-kr8hl\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717872 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-stats-auth\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717888 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-config-volume\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e9895f1f-054d-4b3c-b55d-a02514f51a78-cert\") pod \"ingress-canary-xg824\" (UID: \"e9895f1f-054d-4b3c-b55d-a02514f51a78\") " pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717939 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sxft\" (UniqueName: \"kubernetes.io/projected/a7d810b4-fe88-422e-b3a8-cc2c407039f1-kube-api-access-7sxft\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717969 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.717987 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7d810b4-fe88-422e-b3a8-cc2c407039f1-signing-cabundle\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718006 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rd84\" (UniqueName: \"kubernetes.io/projected/43e33278-4035-4e7a-936c-481775481acf-kube-api-access-9rd84\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718024 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-images\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718044 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-default-certificate\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718070 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d2ab0f-cb2d-41ed-b746-28adc1888942-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/44ad9726-b93f-4dd3-bbdf-b4f6945179a3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7jnfg\" (UID: \"44ad9726-b93f-4dd3-bbdf-b4f6945179a3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpnk6\" (UniqueName: \"kubernetes.io/projected/dc6d7b79-a9de-4462-97a3-2bf2de30f825-kube-api-access-xpnk6\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718140 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5843662e-8938-4478-a6ab-f92c4a2e75d4-serving-cert\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718165 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d222a2b1-b28d-47d4-98bd-cb7b0456c724-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718203 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rctbf\" (UniqueName: \"kubernetes.io/projected/44ad9726-b93f-4dd3-bbdf-b4f6945179a3-kube-api-access-rctbf\") pod \"multus-admission-controller-857f4d67dd-7jnfg\" (UID: \"44ad9726-b93f-4dd3-bbdf-b4f6945179a3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mwqc\" (UniqueName: \"kubernetes.io/projected/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-kube-api-access-6mwqc\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718249 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-metrics-certs\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718267 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59p9r\" (UniqueName: \"kubernetes.io/projected/1680115c-0bb3-4696-8564-a685fe8efdcb-kube-api-access-59p9r\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718293 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dc6d7b79-a9de-4462-97a3-2bf2de30f825-proxy-tls\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718310 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-metrics-tls\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718336 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.718912 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5843662e-8938-4478-a6ab-f92c4a2e75d4-config\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.719575 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.723615 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.723782 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-proxy-tls\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.727048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09517b43-f8f3-4a06-8dbf-0e24e3acc173-apiservice-cert\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.727256 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-registration-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.727383 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5026deaa-f35e-4c06-9c90-cff95d6599a2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6mlnn\" (UID: \"5026deaa-f35e-4c06-9c90-cff95d6599a2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.727417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-csi-data-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.727991 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.728844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1680115c-0bb3-4696-8564-a685fe8efdcb-service-ca-bundle\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.729690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/738dcd07-f77b-4737-8195-07e674320384-config-volume\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.729723 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43e33278-4035-4e7a-936c-481775481acf-srv-cert\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.730288 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d222a2b1-b28d-47d4-98bd-cb7b0456c724-config\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.731515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43e33278-4035-4e7a-936c-481775481acf-profile-collector-cert\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.732235 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-node-bootstrap-token\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.732872 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61924da4-8b4e-4ea9-920b-97252db4a705-config\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.733851 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d222a2b1-b28d-47d4-98bd-cb7b0456c724-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.735020 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a7d810b4-fe88-422e-b3a8-cc2c407039f1-signing-key\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.735627 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-images\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.735960 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d2ab0f-cb2d-41ed-b746-28adc1888942-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.736184 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/eeebe27e-48fb-47b5-bcf4-a640164e4f1e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-57mj8\" (UID: \"eeebe27e-48fb-47b5-bcf4-a640164e4f1e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.736379 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-socket-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.736422 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-plugins-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.736458 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/929c455e-fb87-4885-8612-6888167af1d4-mountpoint-dir\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.738187 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-default-certificate\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.738400 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09517b43-f8f3-4a06-8dbf-0e24e3acc173-webhook-cert\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.738708 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d2ab0f-cb2d-41ed-b746-28adc1888942-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.738938 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/738dcd07-f77b-4737-8195-07e674320384-secret-volume\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.740927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-certs\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.741395 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09517b43-f8f3-4a06-8dbf-0e24e3acc173-tmpfs\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.741555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-stats-auth\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.742472 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-config-volume\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.743947 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dc6d7b79-a9de-4462-97a3-2bf2de30f825-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.744966 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a7d810b4-fe88-422e-b3a8-cc2c407039f1-signing-cabundle\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.745318 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.245299988 +0000 UTC m=+143.283881158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.745337 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e9895f1f-054d-4b3c-b55d-a02514f51a78-cert\") pod \"ingress-canary-xg824\" (UID: \"e9895f1f-054d-4b3c-b55d-a02514f51a78\") " pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.746892 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.747447 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5843662e-8938-4478-a6ab-f92c4a2e75d4-serving-cert\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.748060 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61924da4-8b4e-4ea9-920b-97252db4a705-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.748308 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28ccl\" (UniqueName: \"kubernetes.io/projected/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-kube-api-access-28ccl\") pod \"oauth-openshift-558db77b4-ml4zz\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.749179 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-srv-cert\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.749720 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dc6d7b79-a9de-4462-97a3-2bf2de30f825-proxy-tls\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.751740 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/44ad9726-b93f-4dd3-bbdf-b4f6945179a3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7jnfg\" (UID: \"44ad9726-b93f-4dd3-bbdf-b4f6945179a3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.752870 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.753487 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1680115c-0bb3-4696-8564-a685fe8efdcb-metrics-certs\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.755361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.755455 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.757697 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-metrics-tls\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.762418 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df5bv\" (UniqueName: \"kubernetes.io/projected/6fd89303-90e5-40dc-b9fa-46b4dbd0a805-kube-api-access-df5bv\") pod \"openshift-apiserver-operator-796bbdcf4f-wc4ng\" (UID: \"6fd89303-90e5-40dc-b9fa-46b4dbd0a805\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.763077 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nchb\" (UniqueName: \"kubernetes.io/projected/4eff0e24-4bcf-44be-85a0-15b8311e39bf-kube-api-access-2nchb\") pod \"dns-operator-744455d44c-k8kp9\" (UID: \"4eff0e24-4bcf-44be-85a0-15b8311e39bf\") " pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.778341 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs94d\" (UniqueName: \"kubernetes.io/projected/ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be-kube-api-access-xs94d\") pod \"authentication-operator-69f744f599-s5gjj\" (UID: \"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.781012 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.792517 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.804512 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbqv8\" (UniqueName: \"kubernetes.io/projected/f8c79c29-7251-4b50-8bb1-c066cd8e3060-kube-api-access-fbqv8\") pod \"machine-approver-56656f9798-jk4g4\" (UID: \"f8c79c29-7251-4b50-8bb1-c066cd8e3060\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.818990 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.819593 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.319538979 +0000 UTC m=+143.358120129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.824600 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/0633e0a1-8695-4237-99fc-a7576cc66d15-kube-api-access-h8kw9\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.839125 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hpv5\" (UniqueName: \"kubernetes.io/projected/2830f943-9ae7-428e-8194-35e921cf4d8f-kube-api-access-5hpv5\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.845751 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.852028 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.862813 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.863107 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0633e0a1-8695-4237-99fc-a7576cc66d15-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dtgj9\" (UID: \"0633e0a1-8695-4237-99fc-a7576cc66d15\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.874636 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.880780 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2830f943-9ae7-428e-8194-35e921cf4d8f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8rtl8\" (UID: \"2830f943-9ae7-428e-8194-35e921cf4d8f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.896524 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcdxj\" (UniqueName: \"kubernetes.io/projected/24f2e598-e97c-4583-bc67-cd8afcf6065e-kube-api-access-zcdxj\") pod \"machine-api-operator-5694c8668f-qzh7t\" (UID: \"24f2e598-e97c-4583-bc67-cd8afcf6065e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.909378 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.922397 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: E1007 20:58:12.922739 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.422727245 +0000 UTC m=+143.461308395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.926502 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6gxjd"] Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.928125 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.937188 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.938065 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-bound-sa-token\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.942008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qvcn\" (UniqueName: \"kubernetes.io/projected/6d6919d5-336c-452e-94dd-f37223a44f2c-kube-api-access-6qvcn\") pod \"route-controller-manager-6576b87f9c-85zcb\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.955573 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.958150 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-684cm\" (UniqueName: \"kubernetes.io/projected/21029c24-0c5d-4186-8941-9db6c2d0e31b-kube-api-access-684cm\") pod \"apiserver-7bbb656c7d-t6nd5\" (UID: \"21029c24-0c5d-4186-8941-9db6c2d0e31b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:12 crc kubenswrapper[4926]: I1007 20:58:12.976589 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9rhp\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-kube-api-access-p9rhp\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:12 crc kubenswrapper[4926]: W1007 20:58:12.989357 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64bdf46c_1845_4807_8dfa_4c9526181afc.slice/crio-97d70969d044bd097ac30c344b74974a49239ee32c8763b2b014f02273dbe021 WatchSource:0}: Error finding container 97d70969d044bd097ac30c344b74974a49239ee32c8763b2b014f02273dbe021: Status 404 returned error can't find the container with id 97d70969d044bd097ac30c344b74974a49239ee32c8763b2b014f02273dbe021 Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.035878 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.036770 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.536750838 +0000 UTC m=+143.575331978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.041074 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.049287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd2rw\" (UniqueName: \"kubernetes.io/projected/017aa3d2-9903-4c8a-a6aa-52b06d8c13fb-kube-api-access-vd2rw\") pod \"machine-config-operator-74547568cd-llc7q\" (UID: \"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.050462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplwq\" (UniqueName: \"kubernetes.io/projected/09517b43-f8f3-4a06-8dbf-0e24e3acc173-kube-api-access-gplwq\") pod \"packageserver-d55dfcdfc-d4bs5\" (UID: \"09517b43-f8f3-4a06-8dbf-0e24e3acc173\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.067385 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.082380 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg96q\" (UniqueName: \"kubernetes.io/projected/bd102fa8-8a39-4a85-ac8b-6f16720e4023-kube-api-access-tg96q\") pod \"migrator-59844c95c7-549d6\" (UID: \"bd102fa8-8a39-4a85-ac8b-6f16720e4023\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.088926 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9vst\" (UniqueName: \"kubernetes.io/projected/e9895f1f-054d-4b3c-b55d-a02514f51a78-kube-api-access-j9vst\") pod \"ingress-canary-xg824\" (UID: \"e9895f1f-054d-4b3c-b55d-a02514f51a78\") " pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.104114 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbbwh\" (UniqueName: \"kubernetes.io/projected/bb09fd4d-d43f-4fd5-bf47-97f014ea37ce-kube-api-access-rbbwh\") pod \"machine-config-server-8cwxz\" (UID: \"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce\") " pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.117253 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.123403 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsgjq\" (UniqueName: \"kubernetes.io/projected/8fd9297e-a03a-4bb6-be93-c8b6348a2b50-kube-api-access-vsgjq\") pod \"dns-default-qdvtq\" (UID: \"8fd9297e-a03a-4bb6-be93-c8b6348a2b50\") " pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.126147 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.132651 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k87k7"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.132700 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zq6qq"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.139001 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.139291 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.639278714 +0000 UTC m=+143.677859864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.139893 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61924da4-8b4e-4ea9-920b-97252db4a705-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-smp7q\" (UID: \"61924da4-8b4e-4ea9-920b-97252db4a705\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.156952 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9qcr\" (UniqueName: \"kubernetes.io/projected/929c455e-fb87-4885-8612-6888167af1d4-kube-api-access-s9qcr\") pod \"csi-hostpathplugin-z7z5t\" (UID: \"929c455e-fb87-4885-8612-6888167af1d4\") " pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.161035 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xg824" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.168724 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.175915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8cwxz" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.176813 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwc7l\" (UniqueName: \"kubernetes.io/projected/5026deaa-f35e-4c06-9c90-cff95d6599a2-kube-api-access-nwc7l\") pod \"control-plane-machine-set-operator-78cbb6b69f-6mlnn\" (UID: \"5026deaa-f35e-4c06-9c90-cff95d6599a2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.201718 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa81d6fa-31cf-4c18-a54e-362c1f8780f9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5h9h\" (UID: \"fa81d6fa-31cf-4c18-a54e-362c1f8780f9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.221468 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rd84\" (UniqueName: \"kubernetes.io/projected/43e33278-4035-4e7a-936c-481775481acf-kube-api-access-9rd84\") pod \"catalog-operator-68c6474976-gs2h8\" (UID: \"43e33278-4035-4e7a-936c-481775481acf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.239585 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.239740 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.739714997 +0000 UTC m=+143.778296147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.239893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.240231 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.740216452 +0000 UTC m=+143.778797602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.248028 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.248084 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w6j8g"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.252002 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rctbf\" (UniqueName: \"kubernetes.io/projected/44ad9726-b93f-4dd3-bbdf-b4f6945179a3-kube-api-access-rctbf\") pod \"multus-admission-controller-857f4d67dd-7jnfg\" (UID: \"44ad9726-b93f-4dd3-bbdf-b4f6945179a3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.256721 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.258050 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lgt9\" (UniqueName: \"kubernetes.io/projected/738dcd07-f77b-4737-8195-07e674320384-kube-api-access-6lgt9\") pod \"collect-profiles-29331165-4pss2\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.273655 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.278398 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbrzb\" (UniqueName: \"kubernetes.io/projected/5843662e-8938-4478-a6ab-f92c4a2e75d4-kube-api-access-mbrzb\") pod \"service-ca-operator-777779d784-mktfg\" (UID: \"5843662e-8938-4478-a6ab-f92c4a2e75d4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.300475 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc9dj\" (UniqueName: \"kubernetes.io/projected/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-kube-api-access-cc9dj\") pod \"marketplace-operator-79b997595-5xsvk\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.303521 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.310009 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:13 crc kubenswrapper[4926]: W1007 20:58:13.316300 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaea817a_b104_4b5f_a074_e7196d3ec448.slice/crio-c2d57e6e19c2162bdd06c6a763f34a0a3dcafc3fc5a5d4b035c414a74733d0b5 WatchSource:0}: Error finding container c2d57e6e19c2162bdd06c6a763f34a0a3dcafc3fc5a5d4b035c414a74733d0b5: Status 404 returned error can't find the container with id c2d57e6e19c2162bdd06c6a763f34a0a3dcafc3fc5a5d4b035c414a74733d0b5 Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.316627 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.319774 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr8hl\" (UniqueName: \"kubernetes.io/projected/18d2ab0f-cb2d-41ed-b746-28adc1888942-kube-api-access-kr8hl\") pod \"kube-storage-version-migrator-operator-b67b599dd-j4rcb\" (UID: \"18d2ab0f-cb2d-41ed-b746-28adc1888942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.323661 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.330266 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.338564 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d46d\" (UniqueName: \"kubernetes.io/projected/eeebe27e-48fb-47b5-bcf4-a640164e4f1e-kube-api-access-7d46d\") pod \"package-server-manager-789f6589d5-57mj8\" (UID: \"eeebe27e-48fb-47b5-bcf4-a640164e4f1e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.341952 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.342080 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.842063788 +0000 UTC m=+143.880644938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.342393 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.342691 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.842683187 +0000 UTC m=+143.881264337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.345766 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:13 crc kubenswrapper[4926]: W1007 20:58:13.349399 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf45cac1_ea89_474f_8596_e895d22be327.slice/crio-bebf057a1c7b79eb65e158f8de294924a39477613ddc830fe9dcd2e69a56dc05 WatchSource:0}: Error finding container bebf057a1c7b79eb65e158f8de294924a39477613ddc830fe9dcd2e69a56dc05: Status 404 returned error can't find the container with id bebf057a1c7b79eb65e158f8de294924a39477613ddc830fe9dcd2e69a56dc05 Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.354355 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.372214 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.372464 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpnk6\" (UniqueName: \"kubernetes.io/projected/dc6d7b79-a9de-4462-97a3-2bf2de30f825-kube-api-access-xpnk6\") pod \"machine-config-controller-84d6567774-mdz28\" (UID: \"dc6d7b79-a9de-4462-97a3-2bf2de30f825\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.376771 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.378458 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s5gjj"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.381046 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sxft\" (UniqueName: \"kubernetes.io/projected/a7d810b4-fe88-422e-b3a8-cc2c407039f1-kube-api-access-7sxft\") pod \"service-ca-9c57cc56f-twp9k\" (UID: \"a7d810b4-fe88-422e-b3a8-cc2c407039f1\") " pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.386653 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.395916 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mwqc\" (UniqueName: \"kubernetes.io/projected/8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313-kube-api-access-6mwqc\") pod \"olm-operator-6b444d44fb-sn2tq\" (UID: \"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.404666 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.406132 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n4dfn"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.417530 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.422453 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d222a2b1-b28d-47d4-98bd-cb7b0456c724-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9pjh8\" (UID: \"d222a2b1-b28d-47d4-98bd-cb7b0456c724\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.431734 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.441887 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59p9r\" (UniqueName: \"kubernetes.io/projected/1680115c-0bb3-4696-8564-a685fe8efdcb-kube-api-access-59p9r\") pod \"router-default-5444994796-7928l\" (UID: \"1680115c-0bb3-4696-8564-a685fe8efdcb\") " pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.444412 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.444511 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.944497051 +0000 UTC m=+143.983078201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.444713 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.444966 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:13.944959935 +0000 UTC m=+143.983541085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.452288 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.483041 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.484901 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.492522 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" event={"ID":"b37fc498-33f9-4968-8c6a-1cfb00e123b8","Type":"ContainerStarted","Data":"c93912449079edb50720d1b52345525193e72ad3920202281eeee277e6398686"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.496671 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" event={"ID":"64bdf46c-1845-4807-8dfa-4c9526181afc","Type":"ContainerStarted","Data":"7bbd98f29cd1786c654418d81ee4552e23593a3031be5945147e9e9569d9ac62"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.496703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" event={"ID":"64bdf46c-1845-4807-8dfa-4c9526181afc","Type":"ContainerStarted","Data":"97d70969d044bd097ac30c344b74974a49239ee32c8763b2b014f02273dbe021"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.496716 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.498513 4926 patch_prober.go:28] interesting pod/console-operator-58897d9998-6gxjd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.498565 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" podUID="64bdf46c-1845-4807-8dfa-4c9526181afc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.500023 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8cwxz" event={"ID":"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce","Type":"ContainerStarted","Data":"11bcd66be39a4d46662d9e67bfc13cd12438ebe09d1c46b276718e1e9d5e5107"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.500304 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8cwxz" event={"ID":"bb09fd4d-d43f-4fd5-bf47-97f014ea37ce","Type":"ContainerStarted","Data":"4251b478e4e8e65b2469a672040c74de08b2dba9aeda03319df08203f92a1073"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.500891 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w6j8g" event={"ID":"bf45cac1-ea89-474f-8596-e895d22be327","Type":"ContainerStarted","Data":"bebf057a1c7b79eb65e158f8de294924a39477613ddc830fe9dcd2e69a56dc05"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.507437 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" event={"ID":"87a2e5d7-ad73-4643-b301-5424d27d6e55","Type":"ContainerStarted","Data":"c3195e227e01141ac6582b8788f2bd6461711c45c8f18e72ece1aea4eade2255"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.509182 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" event={"ID":"aaea817a-b104-4b5f-a074-e7196d3ec448","Type":"ContainerStarted","Data":"c2d57e6e19c2162bdd06c6a763f34a0a3dcafc3fc5a5d4b035c414a74733d0b5"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.510643 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" event={"ID":"4d353344-979a-4bb8-9c0c-04c073ca02db","Type":"ContainerStarted","Data":"158adc0cea8bb349c1bb613ed19b1571f264df1ad4f8826afd48479738cdd884"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.514149 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" event={"ID":"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be","Type":"ContainerStarted","Data":"dc019ad6512d82ea922015ca4dffa342743953e166fef3402404b5c8e18550bb"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.523496 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" event={"ID":"f8c79c29-7251-4b50-8bb1-c066cd8e3060","Type":"ContainerStarted","Data":"3e37956887e0fd63c14449188238b13cad00cf9d3cd0d9e4b7656db855743b24"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.523548 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" event={"ID":"f8c79c29-7251-4b50-8bb1-c066cd8e3060","Type":"ContainerStarted","Data":"affb2ae27763b987d8f77cc9bd33c2b5bdfe24a05a19ce6002461f5ad8f0ec2b"} Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.528121 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ggd2q"] Oct 07 20:58:13 crc kubenswrapper[4926]: W1007 20:58:13.540760 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fd89303_90e5_40dc_b9fa_46b4dbd0a805.slice/crio-2e77cdede3e415731b59170fe382bc0aa9479f9fbb733d2cf12dedac1f5b50c6 WatchSource:0}: Error finding container 2e77cdede3e415731b59170fe382bc0aa9479f9fbb733d2cf12dedac1f5b50c6: Status 404 returned error can't find the container with id 2e77cdede3e415731b59170fe382bc0aa9479f9fbb733d2cf12dedac1f5b50c6 Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.546436 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.546663 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.046605734 +0000 UTC m=+144.085186884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.547081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.547635 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.047602475 +0000 UTC m=+144.086183625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.636646 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.647959 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.648244 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.148219403 +0000 UTC m=+144.186800553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: W1007 20:58:13.659977 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a6d69ac_d24d_4e99_9ea8_ffafe15885ec.slice/crio-c98d2abad18ece47103576e3aa9d6828cf47b3f6e4ce1837ff9555f3f0c30a1c WatchSource:0}: Error finding container c98d2abad18ece47103576e3aa9d6828cf47b3f6e4ce1837ff9555f3f0c30a1c: Status 404 returned error can't find the container with id c98d2abad18ece47103576e3aa9d6828cf47b3f6e4ce1837ff9555f3f0c30a1c Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.679395 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.693934 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.721869 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-k8kp9"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.724095 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ml4zz"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.734160 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.750292 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.750625 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.250611915 +0000 UTC m=+144.289193065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.854437 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.854652 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.354635406 +0000 UTC m=+144.393216556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.854766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.855085 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.355077979 +0000 UTC m=+144.393659129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.879224 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.895692 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.899230 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qzh7t"] Oct 07 20:58:13 crc kubenswrapper[4926]: I1007 20:58:13.955444 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:13 crc kubenswrapper[4926]: E1007 20:58:13.955765 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.455751239 +0000 UTC m=+144.494332389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: W1007 20:58:14.040537 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24f2e598_e97c_4583_bc67_cd8afcf6065e.slice/crio-35a21036b62d3018eda5722eea8fee0fc45ff7bd28bf019eb360d2fff8e372ca WatchSource:0}: Error finding container 35a21036b62d3018eda5722eea8fee0fc45ff7bd28bf019eb360d2fff8e372ca: Status 404 returned error can't find the container with id 35a21036b62d3018eda5722eea8fee0fc45ff7bd28bf019eb360d2fff8e372ca Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.056613 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.058306 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.558273475 +0000 UTC m=+144.596854615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.160222 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.160466 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.66043471 +0000 UTC m=+144.699015860 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.161175 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.161584 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.661575944 +0000 UTC m=+144.700157095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.172897 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xg824"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.176123 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.187044 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7jnfg"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.189542 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.195884 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qdvtq"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.198435 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.262002 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.262409 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.762392999 +0000 UTC m=+144.800974149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.321367 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" podStartSLOduration=122.321321428 podStartE2EDuration="2m2.321321428s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:14.314724109 +0000 UTC m=+144.353305259" watchObservedRunningTime="2025-10-07 20:58:14.321321428 +0000 UTC m=+144.359902578" Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.367867 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.369654 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.869593566 +0000 UTC m=+144.908174736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.374080 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.408212 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z7z5t"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.413128 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.421576 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn"] Oct 07 20:58:14 crc kubenswrapper[4926]: W1007 20:58:14.422546 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09517b43_f8f3_4a06_8dbf_0e24e3acc173.slice/crio-89791865ce0a111d640f82e340b94f27716dd7f26ee0725cb7975007c77830c8 WatchSource:0}: Error finding container 89791865ce0a111d640f82e340b94f27716dd7f26ee0725cb7975007c77830c8: Status 404 returned error can't find the container with id 89791865ce0a111d640f82e340b94f27716dd7f26ee0725cb7975007c77830c8 Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.468811 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.469574 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.969523783 +0000 UTC m=+145.008104933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.470051 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.471246 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:14.971229965 +0000 UTC m=+145.009811115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.530609 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5xsvk"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.542680 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mktfg"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.547854 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.547923 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.553315 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.556421 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8cwxz" podStartSLOduration=4.556400187 podStartE2EDuration="4.556400187s" podCreationTimestamp="2025-10-07 20:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:14.553557911 +0000 UTC m=+144.592139061" watchObservedRunningTime="2025-10-07 20:58:14.556400187 +0000 UTC m=+144.594981337" Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.557791 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.573525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" event={"ID":"09517b43-f8f3-4a06-8dbf-0e24e3acc173","Type":"ContainerStarted","Data":"89791865ce0a111d640f82e340b94f27716dd7f26ee0725cb7975007c77830c8"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.574238 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.574459 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.074437622 +0000 UTC m=+145.113018772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.574560 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.575216 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.075181264 +0000 UTC m=+145.113762404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: W1007 20:58:14.579996 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd102fa8_8a39_4a85_ac8b_6f16720e4023.slice/crio-c0cc982c6170c867896fd0b57a80defaddf5b1c6aeda46495b71f4e3099b0041 WatchSource:0}: Error finding container c0cc982c6170c867896fd0b57a80defaddf5b1c6aeda46495b71f4e3099b0041: Status 404 returned error can't find the container with id c0cc982c6170c867896fd0b57a80defaddf5b1c6aeda46495b71f4e3099b0041 Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.580051 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" event={"ID":"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11","Type":"ContainerStarted","Data":"8092dca64f8ef7c434d5afc6dbd7db51b2125a67159219c169d5cc3f6561547a"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.584968 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" event={"ID":"2f2abaf1-1c6b-4768-8228-2cdbf730a457","Type":"ContainerStarted","Data":"c6c9d5cb5743b43bb50b0d05ca5eb19e43e73c77bfc21a072d4cffc738d3ae73"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.587162 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qdvtq" event={"ID":"8fd9297e-a03a-4bb6-be93-c8b6348a2b50","Type":"ContainerStarted","Data":"a296b850b22e200f573f9e9962380723abb2adbe3706314d193a7dee902f71c2"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.599177 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" event={"ID":"2830f943-9ae7-428e-8194-35e921cf4d8f","Type":"ContainerStarted","Data":"3887a0218a8c362ea5efbb202518b9112855232492c25daefff630a4ee1c97ac"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.599246 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" event={"ID":"2830f943-9ae7-428e-8194-35e921cf4d8f","Type":"ContainerStarted","Data":"e9fc598c9303750184925038a8e4d5d408f31b9331ba48d1c6936f66e1c251e3"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.605575 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" event={"ID":"929c455e-fb87-4885-8612-6888167af1d4","Type":"ContainerStarted","Data":"dfb9d99159ca25b4824e3fb6af23414d616884ccfa4fd0bc0a007995296fb804"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.615718 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" event={"ID":"21029c24-0c5d-4186-8941-9db6c2d0e31b","Type":"ContainerStarted","Data":"46e76d724845958ea0f91cd9630feaf04048d8e374e40f97c75c833f5de53cdd"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.622788 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-twp9k"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.646977 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" event={"ID":"5026deaa-f35e-4c06-9c90-cff95d6599a2","Type":"ContainerStarted","Data":"774736b70962b02a4c5039804eb89206f97a308856075449752a7a62662b9e12"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.654486 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" event={"ID":"4d353344-979a-4bb8-9c0c-04c073ca02db","Type":"ContainerStarted","Data":"dd1bf92fbaeb208ae3739c8b5bacd158b5f7cd991dc9bb80c5aad46d2c1463f8"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.656921 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" event={"ID":"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb","Type":"ContainerStarted","Data":"66e73c0bde610044076f98c36b80cdbfde73681690db5978af8cbd018bdbcd05"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.664420 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.677566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.677643 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.177627908 +0000 UTC m=+145.216209058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.678269 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.679510 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.179496014 +0000 UTC m=+145.218077164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.712568 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ggd2q" event={"ID":"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec","Type":"ContainerStarted","Data":"de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.713024 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ggd2q" event={"ID":"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec","Type":"ContainerStarted","Data":"c98d2abad18ece47103576e3aa9d6828cf47b3f6e4ce1837ff9555f3f0c30a1c"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.718356 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" event={"ID":"6fd89303-90e5-40dc-b9fa-46b4dbd0a805","Type":"ContainerStarted","Data":"c012782230b33a80e425b07d97a95c6edd931b2ee769fb17782071247b5f41fa"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.718404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" event={"ID":"6fd89303-90e5-40dc-b9fa-46b4dbd0a805","Type":"ContainerStarted","Data":"2e77cdede3e415731b59170fe382bc0aa9479f9fbb733d2cf12dedac1f5b50c6"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.724284 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" event={"ID":"aaea817a-b104-4b5f-a074-e7196d3ec448","Type":"ContainerStarted","Data":"03d929ce912451ef65424d4be3ec1e4ad4a664801faea96c646ef6aaca79b6b5"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.728492 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" event={"ID":"fa81d6fa-31cf-4c18-a54e-362c1f8780f9","Type":"ContainerStarted","Data":"7fc4fd31d79d5d42808cc7c3e570e31e1f8bb0e8ce1d99e3ba16386f2fbcfe42"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.747463 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.759891 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.766391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" event={"ID":"0633e0a1-8695-4237-99fc-a7576cc66d15","Type":"ContainerStarted","Data":"f5c906c6ad21910acdc76cbb7be873f8042311f9eba8c410bc4f8d1dbdad013c"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.766495 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" event={"ID":"0633e0a1-8695-4237-99fc-a7576cc66d15","Type":"ContainerStarted","Data":"cb788260f10b6d60d83be338193b157bc8013253bd4d06c6c7c4e6f8954dd42a"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.780174 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.781723 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.2816993 +0000 UTC m=+145.320280450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.791387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" event={"ID":"44ad9726-b93f-4dd3-bbdf-b4f6945179a3","Type":"ContainerStarted","Data":"bf77523be190643037f50fa44a7f9952bbda3d5fcd82f1f44242f4b74ba6c98d"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.833187 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb"] Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.836029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" event={"ID":"6d6919d5-336c-452e-94dd-f37223a44f2c","Type":"ContainerStarted","Data":"7cbb8cd844bcd525427c03df9581163f590706f2b2e1e8aceedf04d97d3b354d"} Oct 07 20:58:14 crc kubenswrapper[4926]: W1007 20:58:14.836314 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd222a2b1_b28d_47d4_98bd_cb7b0456c724.slice/crio-ec7a5824c451699db54552165c2ffceb49b3864b54ecc801067ed0a8fa52e7be WatchSource:0}: Error finding container ec7a5824c451699db54552165c2ffceb49b3864b54ecc801067ed0a8fa52e7be: Status 404 returned error can't find the container with id ec7a5824c451699db54552165c2ffceb49b3864b54ecc801067ed0a8fa52e7be Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.838566 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" event={"ID":"4eff0e24-4bcf-44be-85a0-15b8311e39bf","Type":"ContainerStarted","Data":"9f2b192cde78e0763cb042b8adea605c7029b6b3987b0416611a689f9227b87c"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.844677 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w6j8g" event={"ID":"bf45cac1-ea89-474f-8596-e895d22be327","Type":"ContainerStarted","Data":"be4cfb0eb7bba4b7dd79f4b48e8e208122f0ce48274c3a07678b67b0ea2e7378"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.845278 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.846409 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-w6j8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.846444 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w6j8g" podUID="bf45cac1-ea89-474f-8596-e895d22be327" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.848224 4926 generic.go:334] "Generic (PLEG): container finished" podID="87a2e5d7-ad73-4643-b301-5424d27d6e55" containerID="a4dc41c96a9de13fb8c2d391c8bc7d1fd13aa36659c8fd22cc70c8ff687263fa" exitCode=0 Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.848289 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" event={"ID":"87a2e5d7-ad73-4643-b301-5424d27d6e55","Type":"ContainerDied","Data":"a4dc41c96a9de13fb8c2d391c8bc7d1fd13aa36659c8fd22cc70c8ff687263fa"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.855456 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" event={"ID":"43e33278-4035-4e7a-936c-481775481acf","Type":"ContainerStarted","Data":"545fd67a75da4646b8c0264d67e4ec713560e8b025f6fd122ad1a45aa30e1dd6"} Oct 07 20:58:14 crc kubenswrapper[4926]: W1007 20:58:14.876433 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18d2ab0f_cb2d_41ed_b746_28adc1888942.slice/crio-57ed7bc33b45f33b75213ec0c45a8ab3c1fd3e0b7ca6b8258d98a32f46f3dc1b WatchSource:0}: Error finding container 57ed7bc33b45f33b75213ec0c45a8ab3c1fd3e0b7ca6b8258d98a32f46f3dc1b: Status 404 returned error can't find the container with id 57ed7bc33b45f33b75213ec0c45a8ab3c1fd3e0b7ca6b8258d98a32f46f3dc1b Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.881768 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.882560 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.382546805 +0000 UTC m=+145.421127955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.887071 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xg824" event={"ID":"e9895f1f-054d-4b3c-b55d-a02514f51a78","Type":"ContainerStarted","Data":"c4ea3eda51adba55eb3d0ec6f8471696a186df440d7bd7da43ade07f5e57c41d"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.900476 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" event={"ID":"b37fc498-33f9-4968-8c6a-1cfb00e123b8","Type":"ContainerStarted","Data":"2de8240ecf0a5bb884042a1f01bf268880724f81a19442615eac4f9f323a6c1e"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.900753 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.903890 4926 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-k87k7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.903955 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" podUID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.905350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" event={"ID":"ca88c9fa-eafc-4d3c-83ac-e26aaa8fc7be","Type":"ContainerStarted","Data":"b432023ea3cd112739d9a15b25bcece26a586d9b796755d628b17c8f6ff8b2aa"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.918724 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" event={"ID":"f8c79c29-7251-4b50-8bb1-c066cd8e3060","Type":"ContainerStarted","Data":"5c0ce02ed70292e024e3e0a5e5f7d2e0a8b6f9b88132da856bff9816f8767305"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.929176 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7928l" event={"ID":"1680115c-0bb3-4696-8564-a685fe8efdcb","Type":"ContainerStarted","Data":"b5346502b25b9253be325bf6b75dc760df4c00b3e0af1bd87920601eeb9e882b"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.931775 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" event={"ID":"24f2e598-e97c-4583-bc67-cd8afcf6065e","Type":"ContainerStarted","Data":"35a21036b62d3018eda5722eea8fee0fc45ff7bd28bf019eb360d2fff8e372ca"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.935543 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" event={"ID":"492b14c9-39b0-4073-ac89-7dad77ecb8be","Type":"ContainerStarted","Data":"b6e1cdf38f182adeda07c55fc2b412465ad73f8d9b74b41cf0b4db9a8275d632"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.935757 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" event={"ID":"492b14c9-39b0-4073-ac89-7dad77ecb8be","Type":"ContainerStarted","Data":"050fc8e9e30451417566410b0ed891d18344e3ceb6fac0d152cbdda7303e51dd"} Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.982256 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.982513 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.482492174 +0000 UTC m=+145.521073324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:14 crc kubenswrapper[4926]: I1007 20:58:14.983229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:14 crc kubenswrapper[4926]: E1007 20:58:14.986586 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.486567617 +0000 UTC m=+145.525148767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.083718 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.083790 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.583774352 +0000 UTC m=+145.622355502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.083900 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.084141 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.584133573 +0000 UTC m=+145.622714723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.104358 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-6gxjd" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.186950 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.187396 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.68738281 +0000 UTC m=+145.725963960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.290162 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.290546 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.790529224 +0000 UTC m=+145.829110374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.325262 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7928l" podStartSLOduration=123.325213742 podStartE2EDuration="2m3.325213742s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.324716067 +0000 UTC m=+145.363297217" watchObservedRunningTime="2025-10-07 20:58:15.325213742 +0000 UTC m=+145.363794892" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.397946 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.398247 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.898215966 +0000 UTC m=+145.936797116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.398838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.413671 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:15.913649582 +0000 UTC m=+145.952230732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.415363 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dtgj9" podStartSLOduration=123.415347093 podStartE2EDuration="2m3.415347093s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.411284901 +0000 UTC m=+145.449866051" watchObservedRunningTime="2025-10-07 20:58:15.415347093 +0000 UTC m=+145.453928243" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.473498 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-zq6qq" podStartSLOduration=123.473466998 podStartE2EDuration="2m3.473466998s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.436603035 +0000 UTC m=+145.475184195" watchObservedRunningTime="2025-10-07 20:58:15.473466998 +0000 UTC m=+145.512048148" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.474893 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jk4g4" podStartSLOduration=124.474886641 podStartE2EDuration="2m4.474886641s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.472134038 +0000 UTC m=+145.510715188" watchObservedRunningTime="2025-10-07 20:58:15.474886641 +0000 UTC m=+145.513467791" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.500057 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.500417 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.000387191 +0000 UTC m=+146.038968341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.500559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.501014 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.00099599 +0000 UTC m=+146.039577140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.516021 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-w6j8g" podStartSLOduration=123.516001063 podStartE2EDuration="2m3.516001063s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.514777596 +0000 UTC m=+145.553358736" watchObservedRunningTime="2025-10-07 20:58:15.516001063 +0000 UTC m=+145.554582213" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.597754 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ggd2q" podStartSLOduration=123.597734801 podStartE2EDuration="2m3.597734801s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.558338621 +0000 UTC m=+145.596919761" watchObservedRunningTime="2025-10-07 20:58:15.597734801 +0000 UTC m=+145.636315951" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.599098 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-s5gjj" podStartSLOduration=124.599092632 podStartE2EDuration="2m4.599092632s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.597193025 +0000 UTC m=+145.635774175" watchObservedRunningTime="2025-10-07 20:58:15.599092632 +0000 UTC m=+145.637673782" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.602889 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.603608 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.103587998 +0000 UTC m=+146.142169148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.633306 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wc4ng" podStartSLOduration=124.633291885 podStartE2EDuration="2m4.633291885s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.632168741 +0000 UTC m=+145.670749911" watchObservedRunningTime="2025-10-07 20:58:15.633291885 +0000 UTC m=+145.671873035" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.678892 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9l6nb" podStartSLOduration=123.678869301 podStartE2EDuration="2m3.678869301s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.678401487 +0000 UTC m=+145.716982647" watchObservedRunningTime="2025-10-07 20:58:15.678869301 +0000 UTC m=+145.717450461" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.681271 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.690643 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.690706 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.709492 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.709843 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.209822036 +0000 UTC m=+146.248403196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.716434 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" podStartSLOduration=123.716417525 podStartE2EDuration="2m3.716417525s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:15.713631831 +0000 UTC m=+145.752212971" watchObservedRunningTime="2025-10-07 20:58:15.716417525 +0000 UTC m=+145.754998675" Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.810193 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.810379 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.310351431 +0000 UTC m=+146.348932581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.810877 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.811294 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.311278779 +0000 UTC m=+146.349859929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.911711 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:15 crc kubenswrapper[4926]: E1007 20:58:15.911917 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.411898258 +0000 UTC m=+146.450479408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.979177 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" event={"ID":"18d2ab0f-cb2d-41ed-b746-28adc1888942","Type":"ContainerStarted","Data":"57ed7bc33b45f33b75213ec0c45a8ab3c1fd3e0b7ca6b8258d98a32f46f3dc1b"} Oct 07 20:58:15 crc kubenswrapper[4926]: I1007 20:58:15.985303 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" event={"ID":"492b14c9-39b0-4073-ac89-7dad77ecb8be","Type":"ContainerStarted","Data":"01818c26887ca830bbb21d0398ebbcdc69498ea5f09d5a9adec85b6c1724c627"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.005872 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" event={"ID":"43e33278-4035-4e7a-936c-481775481acf","Type":"ContainerStarted","Data":"0641bcc647c30c51edf144c055ba9d03f332bf2229acb6327cc321b0d1a93fff"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.006207 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.012672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.015000 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.51498368 +0000 UTC m=+146.553564830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.015306 4926 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gs2h8 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.015349 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" podUID="43e33278-4035-4e7a-936c-481775481acf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.018763 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-465q2" podStartSLOduration=124.018734304 podStartE2EDuration="2m4.018734304s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.015874297 +0000 UTC m=+146.054455457" watchObservedRunningTime="2025-10-07 20:58:16.018734304 +0000 UTC m=+146.057315454" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.024479 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" event={"ID":"5026deaa-f35e-4c06-9c90-cff95d6599a2","Type":"ContainerStarted","Data":"ea1aca25ad9b3699214106de5c122e0913d98f9c6a444578603ecec9ef08a360"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.043780 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" event={"ID":"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0","Type":"ContainerStarted","Data":"f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.044439 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" event={"ID":"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0","Type":"ContainerStarted","Data":"654c5877f62be05ff6a6661504a41a8f0486514aa3888022962bf70dac18bfa4"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.044516 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.053553 4926 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5xsvk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.053614 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.056358 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" podStartSLOduration=124.056344839 podStartE2EDuration="2m4.056344839s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.053813663 +0000 UTC m=+146.092394813" watchObservedRunningTime="2025-10-07 20:58:16.056344839 +0000 UTC m=+146.094925989" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.063500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" event={"ID":"61924da4-8b4e-4ea9-920b-97252db4a705","Type":"ContainerStarted","Data":"345866a184148c416d37c2ced04a43057df452594d2276745d81e15aefac5c96"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.073161 4926 generic.go:334] "Generic (PLEG): container finished" podID="21029c24-0c5d-4186-8941-9db6c2d0e31b" containerID="8d46033b7b45aff9dab9532b9d87b2047f11e34d8f015dcb9e9cd0d46dda07e6" exitCode=0 Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.073284 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" event={"ID":"21029c24-0c5d-4186-8941-9db6c2d0e31b","Type":"ContainerDied","Data":"8d46033b7b45aff9dab9532b9d87b2047f11e34d8f015dcb9e9cd0d46dda07e6"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.074626 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f2abaf1-1c6b-4768-8228-2cdbf730a457" containerID="529757753b1863fa6bb488e408fd52ad13ba76e246cbaf6e3682e884a8294062" exitCode=0 Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.074683 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" event={"ID":"2f2abaf1-1c6b-4768-8228-2cdbf730a457","Type":"ContainerDied","Data":"529757753b1863fa6bb488e408fd52ad13ba76e246cbaf6e3682e884a8294062"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.083992 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" podStartSLOduration=124.083979544 podStartE2EDuration="2m4.083979544s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.082942223 +0000 UTC m=+146.121523373" watchObservedRunningTime="2025-10-07 20:58:16.083979544 +0000 UTC m=+146.122560694" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.086578 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" event={"ID":"2830f943-9ae7-428e-8194-35e921cf4d8f","Type":"ContainerStarted","Data":"c861a1c84c841afb36afcff938d7b6e0c202ed1a46dd95a44fc7c260b75ab643"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.090818 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" event={"ID":"dc6d7b79-a9de-4462-97a3-2bf2de30f825","Type":"ContainerStarted","Data":"bf532fb7836ead70b1cabebb11b5c2cd821226573a2ef4582b0f902d9235fdb7"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.090852 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" event={"ID":"dc6d7b79-a9de-4462-97a3-2bf2de30f825","Type":"ContainerStarted","Data":"cec2c3639a90c0f6424afec4b3170f70009f158cfc7c0caab4f731167e264e73"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.111111 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6mlnn" podStartSLOduration=124.111093603 podStartE2EDuration="2m4.111093603s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.108524155 +0000 UTC m=+146.147105305" watchObservedRunningTime="2025-10-07 20:58:16.111093603 +0000 UTC m=+146.149674743" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.114531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.114971 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.614949049 +0000 UTC m=+146.653530199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.141839 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8rtl8" podStartSLOduration=124.141823621 podStartE2EDuration="2m4.141823621s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.139949314 +0000 UTC m=+146.178530464" watchObservedRunningTime="2025-10-07 20:58:16.141823621 +0000 UTC m=+146.180404771" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.216062 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.224150 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.724119546 +0000 UTC m=+146.762700696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.245869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" event={"ID":"4eff0e24-4bcf-44be-85a0-15b8311e39bf","Type":"ContainerStarted","Data":"bb677abbdf0116a40fc35cb8d57a65762eafe44c1752155bc0cd0bb5eeb13fed"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.254863 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xg824" event={"ID":"e9895f1f-054d-4b3c-b55d-a02514f51a78","Type":"ContainerStarted","Data":"c506efbedb437ab0faf74dde1e5936d6d1e74fad246ed763b8890dbdc1b3de6f"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.291557 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" podStartSLOduration=124.291521351 podStartE2EDuration="2m4.291521351s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.288538891 +0000 UTC m=+146.327120041" watchObservedRunningTime="2025-10-07 20:58:16.291521351 +0000 UTC m=+146.330102501" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.322785 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.323079 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.823050033 +0000 UTC m=+146.861631183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.370565 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" event={"ID":"738dcd07-f77b-4737-8195-07e674320384","Type":"ContainerStarted","Data":"3d56a82b0b92a09fef99b6885f2f82b0534585cd699f41167dee81b33eae30d4"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.403448 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7928l" event={"ID":"1680115c-0bb3-4696-8564-a685fe8efdcb","Type":"ContainerStarted","Data":"6d6c319c1d3fd526ee2372d2b75e1301e65f4ac4fbcd291aa81c749efaea1254"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.424257 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" event={"ID":"24f2e598-e97c-4583-bc67-cd8afcf6065e","Type":"ContainerStarted","Data":"fb57142e50660d750a6d6f2bc20d8e957e755f68e57ba8faaa3b971ab750babf"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.430975 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" event={"ID":"6d6919d5-336c-452e-94dd-f37223a44f2c","Type":"ContainerStarted","Data":"fb95c0f4ed1744704088e003607b31297eb4fa254a5ab45d7893d9b012ea7c8a"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.431232 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.445456 4926 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-85zcb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.445501 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" podUID="6d6919d5-336c-452e-94dd-f37223a44f2c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.448569 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qdvtq" event={"ID":"8fd9297e-a03a-4bb6-be93-c8b6348a2b50","Type":"ContainerStarted","Data":"5b7006da6225bf0b10845bd6921c94504ce9c9b441c9e734b4c4cb461d01b17d"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.458910 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.459213 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:16.959185954 +0000 UTC m=+146.997767094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.461133 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-xg824" podStartSLOduration=6.461116362 podStartE2EDuration="6.461116362s" podCreationTimestamp="2025-10-07 20:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.354770411 +0000 UTC m=+146.393351561" watchObservedRunningTime="2025-10-07 20:58:16.461116362 +0000 UTC m=+146.499697512" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.474112 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" event={"ID":"bd102fa8-8a39-4a85-ac8b-6f16720e4023","Type":"ContainerStarted","Data":"4902515289a26a1860ec85a3d44804c4bd7c7ab359f92b0f35bbfbe72098b67e"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.474156 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" event={"ID":"bd102fa8-8a39-4a85-ac8b-6f16720e4023","Type":"ContainerStarted","Data":"c0cc982c6170c867896fd0b57a80defaddf5b1c6aeda46495b71f4e3099b0041"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.479408 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" event={"ID":"09517b43-f8f3-4a06-8dbf-0e24e3acc173","Type":"ContainerStarted","Data":"716e22439a1abe2c13f3b344f9685efc22af32f79a39eac8734928b56b8d511c"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.479737 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.485622 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" event={"ID":"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb","Type":"ContainerStarted","Data":"9f414a33ede61d7fe474bcb7d0d797e5146bf209569a333638bbed2493008992"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.485661 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" event={"ID":"017aa3d2-9903-4c8a-a6aa-52b06d8c13fb","Type":"ContainerStarted","Data":"e2d4a38b4411c82d7319a0f265446b609421a96c14d33822c63c90e16fc12669"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.491448 4926 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d4bs5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.491492 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" podUID="09517b43-f8f3-4a06-8dbf-0e24e3acc173" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.492227 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" event={"ID":"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11","Type":"ContainerStarted","Data":"8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.492854 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.496826 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" event={"ID":"a7d810b4-fe88-422e-b3a8-cc2c407039f1","Type":"ContainerStarted","Data":"3251cde1d8c2ac8ab33a225fa0d6d1add3f39a0610eebf50b3b406337e5e2ebe"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.496853 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" event={"ID":"a7d810b4-fe88-422e-b3a8-cc2c407039f1","Type":"ContainerStarted","Data":"fec55ce3444cb844886dbcecd4c8a72e51aef491c01a0e370d954ef96a424a5a"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.500171 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" event={"ID":"44ad9726-b93f-4dd3-bbdf-b4f6945179a3","Type":"ContainerStarted","Data":"a26b2f93dc9c2ed6a5bb100af3b2a806780930dd84c1258e2659ac73ea855b6a"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.502780 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" podStartSLOduration=124.50277045 podStartE2EDuration="2m4.50277045s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.462729181 +0000 UTC m=+146.501310351" watchObservedRunningTime="2025-10-07 20:58:16.50277045 +0000 UTC m=+146.541351600" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.508032 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" event={"ID":"eeebe27e-48fb-47b5-bcf4-a640164e4f1e","Type":"ContainerStarted","Data":"404d850e7065135b30ffa647071c6e2db1c367bbad1d43aaaef1095e9a0f78ee"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.546164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" event={"ID":"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313","Type":"ContainerStarted","Data":"12f1d4c7df7048a4ea714e8b3b4d2d4b9cb56740cfd307d57c354eeed3a927fa"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.547047 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.568187 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.569814 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.069798014 +0000 UTC m=+147.108379164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.571260 4926 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sn2tq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.571292 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" podUID="8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.573447 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" podStartSLOduration=124.573437714 podStartE2EDuration="2m4.573437714s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.505114511 +0000 UTC m=+146.543695661" watchObservedRunningTime="2025-10-07 20:58:16.573437714 +0000 UTC m=+146.612018864" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.575358 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-twp9k" podStartSLOduration=124.575351532 podStartE2EDuration="2m4.575351532s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.536530129 +0000 UTC m=+146.575111279" watchObservedRunningTime="2025-10-07 20:58:16.575351532 +0000 UTC m=+146.613932682" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.578635 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" event={"ID":"d222a2b1-b28d-47d4-98bd-cb7b0456c724","Type":"ContainerStarted","Data":"ec7a5824c451699db54552165c2ffceb49b3864b54ecc801067ed0a8fa52e7be"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.614493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" event={"ID":"5843662e-8938-4478-a6ab-f92c4a2e75d4","Type":"ContainerStarted","Data":"c04b140e2aa273c4f54e0e5432cde283c4c6c5ae7a8a1e45a24e3388a72edbd8"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.626164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" event={"ID":"fa81d6fa-31cf-4c18-a54e-362c1f8780f9","Type":"ContainerStarted","Data":"83baf1b680fab295b8ef754abac3af93461d895f3ce5c5649476543f6e930ce8"} Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.627096 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-w6j8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.627138 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w6j8g" podUID="bf45cac1-ea89-474f-8596-e895d22be327" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.662999 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.670174 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.673043 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.173029111 +0000 UTC m=+147.211610261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.693514 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:16 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:16 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:16 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.693570 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.734565 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-llc7q" podStartSLOduration=124.734545579 podStartE2EDuration="2m4.734545579s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.733834817 +0000 UTC m=+146.772415967" watchObservedRunningTime="2025-10-07 20:58:16.734545579 +0000 UTC m=+146.773126729" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.771898 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.774882 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.274852026 +0000 UTC m=+147.313433176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.775321 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.776882 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" podStartSLOduration=125.776868857 podStartE2EDuration="2m5.776868857s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.774952949 +0000 UTC m=+146.813534099" watchObservedRunningTime="2025-10-07 20:58:16.776868857 +0000 UTC m=+146.815450007" Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.779369 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.279354872 +0000 UTC m=+147.317936042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.792517 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.867744 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" podStartSLOduration=124.867731471 podStartE2EDuration="2m4.867731471s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.865953867 +0000 UTC m=+146.904535017" watchObservedRunningTime="2025-10-07 20:58:16.867731471 +0000 UTC m=+146.906312621" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.878414 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.878970 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.378952569 +0000 UTC m=+147.417533729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.916825 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5h9h" podStartSLOduration=124.916808423 podStartE2EDuration="2m4.916808423s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.915132982 +0000 UTC m=+146.953714132" watchObservedRunningTime="2025-10-07 20:58:16.916808423 +0000 UTC m=+146.955389573" Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.979863 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:16 crc kubenswrapper[4926]: E1007 20:58:16.980265 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.480249918 +0000 UTC m=+147.518831058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:16 crc kubenswrapper[4926]: I1007 20:58:16.993926 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" podStartSLOduration=124.993907841 podStartE2EDuration="2m4.993907841s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:16.94751587 +0000 UTC m=+146.986097020" watchObservedRunningTime="2025-10-07 20:58:16.993907841 +0000 UTC m=+147.032488991" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.082658 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.083301 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.58328643 +0000 UTC m=+147.621867580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.184475 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.184854 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.684839586 +0000 UTC m=+147.723420726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.285641 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.286050 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.786035692 +0000 UTC m=+147.824616842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.387697 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.388067 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.888050863 +0000 UTC m=+147.926632013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.489263 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.489445 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.989420474 +0000 UTC m=+148.028001614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.489528 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.489852 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:17.989840866 +0000 UTC m=+148.028422016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.590675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.591017 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.09097909 +0000 UTC m=+148.129560230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.649328 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" event={"ID":"738dcd07-f77b-4737-8195-07e674320384","Type":"ContainerStarted","Data":"ba557e2ac9cdc96297d3c0fc084ef707e8ec2d9d96224aea6096ee318ca53b9b"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.650820 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" event={"ID":"8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313","Type":"ContainerStarted","Data":"d11d2db8559bf6a5fe809e66d4eb3133199fb437ec32dd9400baba17c090b7e2"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.651316 4926 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sn2tq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.651361 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" podUID="8f5d2f0a-c6cd-4e94-a3ca-e5207bdac313" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.652294 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" event={"ID":"87a2e5d7-ad73-4643-b301-5424d27d6e55","Type":"ContainerStarted","Data":"25f92e88ae3b31bb7adec476abd18ca3e76954ad6fa38bb816412f0432cc1eeb"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.652405 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.653375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" event={"ID":"dc6d7b79-a9de-4462-97a3-2bf2de30f825","Type":"ContainerStarted","Data":"3b0419efe112a67f62a07c98eb80171c0b8f11f597c7fa1d20c68a4db43bd658"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.654674 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mktfg" event={"ID":"5843662e-8938-4478-a6ab-f92c4a2e75d4","Type":"ContainerStarted","Data":"e5396a6b4842575dcb19abd373f9ced564d21b5020795b36e8e5117441ce3e28"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.673934 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-k8kp9" event={"ID":"4eff0e24-4bcf-44be-85a0-15b8311e39bf","Type":"ContainerStarted","Data":"17e174d114f2a3cccd64d6afcf8bb3bd3fa9c1c93ab027cfd30a590054835cd1"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.681663 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" event={"ID":"18d2ab0f-cb2d-41ed-b746-28adc1888942","Type":"ContainerStarted","Data":"b6ab938804760ded677f0bc336ca7505d5e9a92a7706b47f8394a60778f1e9d0"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.683941 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" event={"ID":"61924da4-8b4e-4ea9-920b-97252db4a705","Type":"ContainerStarted","Data":"1f0592f0d795741e2b8a998c6e4c5238c9580778bc619e82b490fe5bbc1d9012"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.685504 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" event={"ID":"eeebe27e-48fb-47b5-bcf4-a640164e4f1e","Type":"ContainerStarted","Data":"75dbcb79bf72634ba95256f9bcd14f88a867da92cd95a6872f38defbaa78922a"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.685527 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" event={"ID":"eeebe27e-48fb-47b5-bcf4-a640164e4f1e","Type":"ContainerStarted","Data":"d7f5078d863c64ea128896e2858b56f234f8ac4244ada50d3d7574b28b014cfb"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.685878 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.691495 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:17 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:17 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:17 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.691561 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.691782 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" event={"ID":"21029c24-0c5d-4186-8941-9db6c2d0e31b","Type":"ContainerStarted","Data":"7d164e68e01ccd4c4216bdbbace807295c85821c37d6d30d0a400f0e54750b0f"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.692703 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.693045 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.193031172 +0000 UTC m=+148.231612312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.700614 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" event={"ID":"d222a2b1-b28d-47d4-98bd-cb7b0456c724","Type":"ContainerStarted","Data":"12bff506bf25fc4897e40087695ff64d7675c8e53f16e37bf0fc563795bb544b"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.710883 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qdvtq" event={"ID":"8fd9297e-a03a-4bb6-be93-c8b6348a2b50","Type":"ContainerStarted","Data":"f55da61523d1005cc1a036c79679c80bb5b3bd7d78fa536d783fdd9967713d89"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.711033 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.716485 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" event={"ID":"929c455e-fb87-4885-8612-6888167af1d4","Type":"ContainerStarted","Data":"9ec5c51d97b3f3aba25bc68c5c3fd3290a0ffe252660c3c7c57162eb9e6c4d08"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.717661 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" event={"ID":"bd102fa8-8a39-4a85-ac8b-6f16720e4023","Type":"ContainerStarted","Data":"864109f4897da634caa1f4b5c69acf2a5fc6d0f7d65f11807076186248bae042"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.719391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" event={"ID":"24f2e598-e97c-4583-bc67-cd8afcf6065e","Type":"ContainerStarted","Data":"6f2f3efd16fd7d37bb4ef908a677ee09391dac1e86141e6380c4613357b7d632"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.721006 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" event={"ID":"44ad9726-b93f-4dd3-bbdf-b4f6945179a3","Type":"ContainerStarted","Data":"93af30267c9791f1e9d20c89e06842c48bc196f5e54eb2151f9007c354f1cea0"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.733679 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" event={"ID":"2f2abaf1-1c6b-4768-8228-2cdbf730a457","Type":"ContainerStarted","Data":"39a597081c2077aa8c276172b10d988e0f6272fc55525b420808b8b0be6044d6"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.733721 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" event={"ID":"2f2abaf1-1c6b-4768-8228-2cdbf730a457","Type":"ContainerStarted","Data":"e9ae17ba020d4d08864ddff32ef82705e0db0e67855cba0bbb863819ffd9b859"} Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.736320 4926 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5xsvk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.736363 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.747507 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gs2h8" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.752464 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" podStartSLOduration=125.752446806 podStartE2EDuration="2m5.752446806s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:17.708688035 +0000 UTC m=+147.747269185" watchObservedRunningTime="2025-10-07 20:58:17.752446806 +0000 UTC m=+147.791027956" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.754582 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.781822 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.782313 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.782666 4926 patch_prober.go:28] interesting pod/apiserver-76f77b778f-n4dfn container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.782716 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" podUID="2f2abaf1-1c6b-4768-8228-2cdbf730a457" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.786924 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" podStartSLOduration=125.786910447 podStartE2EDuration="2m5.786910447s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:17.754577221 +0000 UTC m=+147.793158371" watchObservedRunningTime="2025-10-07 20:58:17.786910447 +0000 UTC m=+147.825491597" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.787333 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-smp7q" podStartSLOduration=125.78732791 podStartE2EDuration="2m5.78732791s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:17.7863508 +0000 UTC m=+147.824931950" watchObservedRunningTime="2025-10-07 20:58:17.78732791 +0000 UTC m=+147.825909060" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.795980 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.798793 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.298768875 +0000 UTC m=+148.337350025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.889119 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9pjh8" podStartSLOduration=125.889099673 podStartE2EDuration="2m5.889099673s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:17.850783786 +0000 UTC m=+147.889364926" watchObservedRunningTime="2025-10-07 20:58:17.889099673 +0000 UTC m=+147.927680823" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.900574 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:17 crc kubenswrapper[4926]: E1007 20:58:17.928691 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.428673208 +0000 UTC m=+148.467254408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.956002 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-j4rcb" podStartSLOduration=125.955984832 podStartE2EDuration="2m5.955984832s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:17.896508436 +0000 UTC m=+147.935089586" watchObservedRunningTime="2025-10-07 20:58:17.955984832 +0000 UTC m=+147.994565982" Oct 07 20:58:17 crc kubenswrapper[4926]: I1007 20:58:17.956392 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" podStartSLOduration=125.956385004 podStartE2EDuration="2m5.956385004s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:17.953657122 +0000 UTC m=+147.992238272" watchObservedRunningTime="2025-10-07 20:58:17.956385004 +0000 UTC m=+147.994966144" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.001689 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.001972 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.501957081 +0000 UTC m=+148.540538231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.030357 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" podStartSLOduration=126.030344508 podStartE2EDuration="2m6.030344508s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.029391859 +0000 UTC m=+148.067973009" watchObservedRunningTime="2025-10-07 20:58:18.030344508 +0000 UTC m=+148.068925658" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.038539 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d4bs5" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.041518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.041541 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.045285 4926 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-t6nd5 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.045343 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" podUID="21029c24-0c5d-4186-8941-9db6c2d0e31b" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.103301 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.103656 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.603644771 +0000 UTC m=+148.642225921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.134018 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mdz28" podStartSLOduration=126.134002878 podStartE2EDuration="2m6.134002878s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.084365859 +0000 UTC m=+148.122947009" watchObservedRunningTime="2025-10-07 20:58:18.134002878 +0000 UTC m=+148.172584028" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.205743 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.206403 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.706389434 +0000 UTC m=+148.744970584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.216067 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" podStartSLOduration=127.216053376 podStartE2EDuration="2m7.216053376s" podCreationTimestamp="2025-10-07 20:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.215827359 +0000 UTC m=+148.254408509" watchObservedRunningTime="2025-10-07 20:58:18.216053376 +0000 UTC m=+148.254634526" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.255688 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-549d6" podStartSLOduration=126.255670362 podStartE2EDuration="2m6.255670362s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.250542177 +0000 UTC m=+148.289123317" watchObservedRunningTime="2025-10-07 20:58:18.255670362 +0000 UTC m=+148.294251512" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.333571 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.334184 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.834167192 +0000 UTC m=+148.872748342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.354057 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qdvtq" podStartSLOduration=8.354030772 podStartE2EDuration="8.354030772s" podCreationTimestamp="2025-10-07 20:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.321018225 +0000 UTC m=+148.359599375" watchObservedRunningTime="2025-10-07 20:58:18.354030772 +0000 UTC m=+148.392611922" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.390779 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7jnfg" podStartSLOduration=126.390761001 podStartE2EDuration="2m6.390761001s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.359126336 +0000 UTC m=+148.397707476" watchObservedRunningTime="2025-10-07 20:58:18.390761001 +0000 UTC m=+148.429342151" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.428060 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qzh7t" podStartSLOduration=126.428038637 podStartE2EDuration="2m6.428038637s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:18.423772788 +0000 UTC m=+148.462353938" watchObservedRunningTime="2025-10-07 20:58:18.428038637 +0000 UTC m=+148.466619787" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.435896 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.436462 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:18.936438391 +0000 UTC m=+148.975019541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.537651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.538100 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.03808483 +0000 UTC m=+149.076665980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.638469 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.638779 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.638817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.638851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.638880 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.644056 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.144026599 +0000 UTC m=+149.182607749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.644171 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.651754 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.664262 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.686042 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.692329 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:18 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:18 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:18 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.692383 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.707865 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.747443 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.747915 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.247898645 +0000 UTC m=+149.286479795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.791645 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" event={"ID":"929c455e-fb87-4885-8612-6888167af1d4","Type":"ContainerStarted","Data":"ab3ad2c1d4b7f17e2d13adf45c7b01de346be0c9fc55515446258a4cd4d97a46"} Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.791717 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" event={"ID":"929c455e-fb87-4885-8612-6888167af1d4","Type":"ContainerStarted","Data":"0ef8253b9001116f5d8c5309ba9b6964b2cafd5cefb6c16e0197515f44422ffc"} Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.802448 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.805327 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sn2tq" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.815393 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.849937 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.850384 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.350353099 +0000 UTC m=+149.388934249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.946839 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-29lbg"] Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.948049 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.957092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.964006 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 20:58:18 crc kubenswrapper[4926]: E1007 20:58:18.975346 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.475328073 +0000 UTC m=+149.513909223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:18 crc kubenswrapper[4926]: I1007 20:58:18.976519 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29lbg"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.059366 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.059555 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkzw5\" (UniqueName: \"kubernetes.io/projected/822a04da-8549-47c1-9994-bcc9b7ca1ed7-kube-api-access-xkzw5\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.059597 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-catalog-content\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.059631 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-utilities\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.059736 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.559722552 +0000 UTC m=+149.598303702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.153071 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbgt9"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.154449 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.161815 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkzw5\" (UniqueName: \"kubernetes.io/projected/822a04da-8549-47c1-9994-bcc9b7ca1ed7-kube-api-access-xkzw5\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.161857 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.161890 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-catalog-content\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.161920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-utilities\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.162483 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.162772 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-utilities\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.163144 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.663134814 +0000 UTC m=+149.701715964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.163184 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-catalog-content\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.213914 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbgt9"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.224300 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkzw5\" (UniqueName: \"kubernetes.io/projected/822a04da-8549-47c1-9994-bcc9b7ca1ed7-kube-api-access-xkzw5\") pod \"community-operators-29lbg\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.269862 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.270051 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tssws\" (UniqueName: \"kubernetes.io/projected/86367f52-f7bb-4644-ac82-c1bda5df5784-kube-api-access-tssws\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.270073 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-utilities\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.270167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-catalog-content\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.270306 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.77029197 +0000 UTC m=+149.808873120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.337488 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.371870 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.371957 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-catalog-content\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.371984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tssws\" (UniqueName: \"kubernetes.io/projected/86367f52-f7bb-4644-ac82-c1bda5df5784-kube-api-access-tssws\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.372001 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-utilities\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.372403 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-utilities\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.372626 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.87261498 +0000 UTC m=+149.911196130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.373081 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-catalog-content\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.397869 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wvl5m"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.398905 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.405915 4926 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.446816 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wvl5m"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.474192 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.474566 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxvms\" (UniqueName: \"kubernetes.io/projected/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-kube-api-access-qxvms\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.474611 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-catalog-content\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.474668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-utilities\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.474771 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:19.974757275 +0000 UTC m=+150.013338425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.475160 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tssws\" (UniqueName: \"kubernetes.io/projected/86367f52-f7bb-4644-ac82-c1bda5df5784-kube-api-access-tssws\") pod \"certified-operators-dbgt9\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.560431 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.579676 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.579739 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-catalog-content\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.579817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-utilities\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.579854 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxvms\" (UniqueName: \"kubernetes.io/projected/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-kube-api-access-qxvms\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.580635 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:20.080609171 +0000 UTC m=+150.119190321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.581319 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-catalog-content\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.581557 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-utilities\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.609765 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hknpx"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.610722 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.638462 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hknpx"] Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.644358 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxvms\" (UniqueName: \"kubernetes.io/projected/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-kube-api-access-qxvms\") pod \"community-operators-wvl5m\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.684056 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.684391 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rjh8\" (UniqueName: \"kubernetes.io/projected/932d000c-28ca-4128-a049-01b7c2960ed1-kube-api-access-4rjh8\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.684458 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-catalog-content\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.684500 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-utilities\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.684692 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:20.184669104 +0000 UTC m=+150.223250254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.691867 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:19 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:19 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:19 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.691933 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:19 crc kubenswrapper[4926]: W1007 20:58:19.735347 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-6a6c24fb0630eb0ff2cf57abb1662ef2027b8d944498a91cc1c448c9ff3430cb WatchSource:0}: Error finding container 6a6c24fb0630eb0ff2cf57abb1662ef2027b8d944498a91cc1c448c9ff3430cb: Status 404 returned error can't find the container with id 6a6c24fb0630eb0ff2cf57abb1662ef2027b8d944498a91cc1c448c9ff3430cb Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.748128 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.788442 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.788893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rjh8\" (UniqueName: \"kubernetes.io/projected/932d000c-28ca-4128-a049-01b7c2960ed1-kube-api-access-4rjh8\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.788947 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:20.288921592 +0000 UTC m=+150.327502742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.789695 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-catalog-content\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.789793 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-utilities\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.790377 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-utilities\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.790587 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-catalog-content\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.837508 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rjh8\" (UniqueName: \"kubernetes.io/projected/932d000c-28ca-4128-a049-01b7c2960ed1-kube-api-access-4rjh8\") pod \"certified-operators-hknpx\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.844125 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.847630 4926 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-07T20:58:19.405940027Z","Handler":null,"Name":""} Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.868391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" event={"ID":"929c455e-fb87-4885-8612-6888167af1d4","Type":"ContainerStarted","Data":"bc5bb5fe0296351626701f9e93f7db318ae9332418393cdc5d25f76ff6792edd"} Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.872819 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"efcd6b9e80608e0e4e5f16cd5a232ac48400a7ea40eedca6b73377d5eeb427ee"} Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.875402 4926 generic.go:334] "Generic (PLEG): container finished" podID="738dcd07-f77b-4737-8195-07e674320384" containerID="ba557e2ac9cdc96297d3c0fc084ef707e8ec2d9d96224aea6096ee318ca53b9b" exitCode=0 Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.875446 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" event={"ID":"738dcd07-f77b-4737-8195-07e674320384","Type":"ContainerDied","Data":"ba557e2ac9cdc96297d3c0fc084ef707e8ec2d9d96224aea6096ee318ca53b9b"} Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.890450 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:19 crc kubenswrapper[4926]: E1007 20:58:19.891208 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 20:58:20.39117776 +0000 UTC m=+150.429758910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.914556 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6a6c24fb0630eb0ff2cf57abb1662ef2027b8d944498a91cc1c448c9ff3430cb"} Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.915506 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-z7z5t" podStartSLOduration=9.915484434 podStartE2EDuration="9.915484434s" podCreationTimestamp="2025-10-07 20:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:19.915226616 +0000 UTC m=+149.953807766" watchObservedRunningTime="2025-10-07 20:58:19.915484434 +0000 UTC m=+149.954065584" Oct 07 20:58:19 crc kubenswrapper[4926]: W1007 20:58:19.956854 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-875d895fabdc4bf6e65ed73b73d3b8e29920ff520475485b1ca6f8accee923d9 WatchSource:0}: Error finding container 875d895fabdc4bf6e65ed73b73d3b8e29920ff520475485b1ca6f8accee923d9: Status 404 returned error can't find the container with id 875d895fabdc4bf6e65ed73b73d3b8e29920ff520475485b1ca6f8accee923d9 Oct 07 20:58:19 crc kubenswrapper[4926]: I1007 20:58:19.992475 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:20 crc kubenswrapper[4926]: E1007 20:58:19.995742 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 20:58:20.495729737 +0000 UTC m=+150.534310887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kfjx5" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.011383 4926 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.011423 4926 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.093106 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.093125 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29lbg"] Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.104603 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.197917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.200717 4926 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.200749 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.209092 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbgt9"] Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.360517 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kfjx5\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.449695 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wvl5m"] Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.464959 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.513379 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hknpx"] Oct 07 20:58:20 crc kubenswrapper[4926]: W1007 20:58:20.553074 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdc058ed_3a5e_4c9c_98ff_6cd72519dd10.slice/crio-8c624bd212a789b1db559307e3adcf1eb5cf3e5046d7827f4e62475c423ba647 WatchSource:0}: Error finding container 8c624bd212a789b1db559307e3adcf1eb5cf3e5046d7827f4e62475c423ba647: Status 404 returned error can't find the container with id 8c624bd212a789b1db559307e3adcf1eb5cf3e5046d7827f4e62475c423ba647 Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.724566 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.733299 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:20 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:20 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:20 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.733383 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.805517 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kfjx5"] Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.924606 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bpwnc"] Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.925560 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.928501 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.936292 4926 generic.go:334] "Generic (PLEG): container finished" podID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerID="dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f" exitCode=0 Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.936395 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerDied","Data":"dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.936427 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerStarted","Data":"8c624bd212a789b1db559307e3adcf1eb5cf3e5046d7827f4e62475c423ba647"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.938247 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.939968 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpwnc"] Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.944794 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"09e121bdf381d096f4943dcf777b0b7c21e0e89a55327ae0df233f13d9f3ddaa"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.944884 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"875d895fabdc4bf6e65ed73b73d3b8e29920ff520475485b1ca6f8accee923d9"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.945041 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.947510 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b8bfe83ef2380bd04254ba7f5210e39b28158faf9143c71b8921b2ee0b3b2c84"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.949923 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" event={"ID":"1ec7685a-e96e-434f-be3a-cda923c04ba1","Type":"ContainerStarted","Data":"c786f57d42b2379303ec14bb07af640a5a553c6b4dd9adb7d9b3596992ebedea"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.954035 4926 generic.go:334] "Generic (PLEG): container finished" podID="932d000c-28ca-4128-a049-01b7c2960ed1" containerID="f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074" exitCode=0 Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.954079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hknpx" event={"ID":"932d000c-28ca-4128-a049-01b7c2960ed1","Type":"ContainerDied","Data":"f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.954096 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hknpx" event={"ID":"932d000c-28ca-4128-a049-01b7c2960ed1","Type":"ContainerStarted","Data":"586f609eb95df2e81f9cbdb952018e729d48a0cb39cd600ff2b8f2746a9d840d"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.960455 4926 generic.go:334] "Generic (PLEG): container finished" podID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerID="4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3" exitCode=0 Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.960557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgt9" event={"ID":"86367f52-f7bb-4644-ac82-c1bda5df5784","Type":"ContainerDied","Data":"4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.960589 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgt9" event={"ID":"86367f52-f7bb-4644-ac82-c1bda5df5784","Type":"ContainerStarted","Data":"d1bd551db3afcd1b1a793aa09933eb79a4ed5ca9152abf29e414fddc3d0ab777"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.965219 4926 generic.go:334] "Generic (PLEG): container finished" podID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerID="5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12" exitCode=0 Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.965280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29lbg" event={"ID":"822a04da-8549-47c1-9994-bcc9b7ca1ed7","Type":"ContainerDied","Data":"5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.965300 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29lbg" event={"ID":"822a04da-8549-47c1-9994-bcc9b7ca1ed7","Type":"ContainerStarted","Data":"2017d0f25c103a8d47f47609b281a22bccfb9a7b6f49b69e4ae33a78ecf11cb3"} Oct 07 20:58:20 crc kubenswrapper[4926]: I1007 20:58:20.966666 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c451be9fcd458b3341f55ebd6ae4f4c6e02d639e83f7f20a5c2e48487f2862cc"} Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.024655 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clc9t\" (UniqueName: \"kubernetes.io/projected/9035f772-a1f8-425d-90a3-00f038a08bf0-kube-api-access-clc9t\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.025069 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-catalog-content\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.025093 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-utilities\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.132862 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-catalog-content\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.132905 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-utilities\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.132985 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clc9t\" (UniqueName: \"kubernetes.io/projected/9035f772-a1f8-425d-90a3-00f038a08bf0-kube-api-access-clc9t\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.134808 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-utilities\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.134836 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-catalog-content\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.151848 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clc9t\" (UniqueName: \"kubernetes.io/projected/9035f772-a1f8-425d-90a3-00f038a08bf0-kube-api-access-clc9t\") pod \"redhat-marketplace-bpwnc\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.184648 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.300771 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.335542 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lgt9\" (UniqueName: \"kubernetes.io/projected/738dcd07-f77b-4737-8195-07e674320384-kube-api-access-6lgt9\") pod \"738dcd07-f77b-4737-8195-07e674320384\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.335592 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/738dcd07-f77b-4737-8195-07e674320384-config-volume\") pod \"738dcd07-f77b-4737-8195-07e674320384\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.335612 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/738dcd07-f77b-4737-8195-07e674320384-secret-volume\") pod \"738dcd07-f77b-4737-8195-07e674320384\" (UID: \"738dcd07-f77b-4737-8195-07e674320384\") " Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.336388 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/738dcd07-f77b-4737-8195-07e674320384-config-volume" (OuterVolumeSpecName: "config-volume") pod "738dcd07-f77b-4737-8195-07e674320384" (UID: "738dcd07-f77b-4737-8195-07e674320384"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.343564 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/738dcd07-f77b-4737-8195-07e674320384-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "738dcd07-f77b-4737-8195-07e674320384" (UID: "738dcd07-f77b-4737-8195-07e674320384"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.343743 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/738dcd07-f77b-4737-8195-07e674320384-kube-api-access-6lgt9" (OuterVolumeSpecName: "kube-api-access-6lgt9") pod "738dcd07-f77b-4737-8195-07e674320384" (UID: "738dcd07-f77b-4737-8195-07e674320384"). InnerVolumeSpecName "kube-api-access-6lgt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.345289 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j7wkx"] Oct 07 20:58:21 crc kubenswrapper[4926]: E1007 20:58:21.345529 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="738dcd07-f77b-4737-8195-07e674320384" containerName="collect-profiles" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.345550 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="738dcd07-f77b-4737-8195-07e674320384" containerName="collect-profiles" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.345700 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="738dcd07-f77b-4737-8195-07e674320384" containerName="collect-profiles" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.346358 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.351909 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7wkx"] Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.437229 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-utilities\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.437527 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glmcp\" (UniqueName: \"kubernetes.io/projected/7742a5ac-dd42-47b0-af53-077cc6553f78-kube-api-access-glmcp\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.437613 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-catalog-content\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.437647 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lgt9\" (UniqueName: \"kubernetes.io/projected/738dcd07-f77b-4737-8195-07e674320384-kube-api-access-6lgt9\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.437658 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/738dcd07-f77b-4737-8195-07e674320384-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.437667 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/738dcd07-f77b-4737-8195-07e674320384-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.523538 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpwnc"] Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.538915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-catalog-content\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.538965 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-utilities\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.538997 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glmcp\" (UniqueName: \"kubernetes.io/projected/7742a5ac-dd42-47b0-af53-077cc6553f78-kube-api-access-glmcp\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.539380 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-catalog-content\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.539548 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-utilities\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.559461 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glmcp\" (UniqueName: \"kubernetes.io/projected/7742a5ac-dd42-47b0-af53-077cc6553f78-kube-api-access-glmcp\") pod \"redhat-marketplace-j7wkx\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.600688 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mq6kz" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.669756 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.683752 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:21 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:21 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:21 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.683883 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.883387 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7wkx"] Oct 07 20:58:21 crc kubenswrapper[4926]: E1007 20:58:21.891481 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9035f772_a1f8_425d_90a3_00f038a08bf0.slice/crio-conmon-10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9035f772_a1f8_425d_90a3_00f038a08bf0.slice/crio-10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70.scope\": RecentStats: unable to find data in memory cache]" Oct 07 20:58:21 crc kubenswrapper[4926]: I1007 20:58:21.997649 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7wkx" event={"ID":"7742a5ac-dd42-47b0-af53-077cc6553f78","Type":"ContainerStarted","Data":"a0cda9d2a91bb86cc4be854118dc1c7a318d11e5e9756e744b9e71f8e47fa39e"} Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.008089 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" event={"ID":"1ec7685a-e96e-434f-be3a-cda923c04ba1","Type":"ContainerStarted","Data":"ea2bf0b3bc9707d468947c4f6ab9b0fbfa3a88740a287d98991f49dd0364ffeb"} Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.008260 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.013866 4926 generic.go:334] "Generic (PLEG): container finished" podID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerID="10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70" exitCode=0 Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.014183 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpwnc" event={"ID":"9035f772-a1f8-425d-90a3-00f038a08bf0","Type":"ContainerDied","Data":"10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70"} Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.014259 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpwnc" event={"ID":"9035f772-a1f8-425d-90a3-00f038a08bf0","Type":"ContainerStarted","Data":"8505377e1c8a44ea4fccc54e036ae01f82a45f0e579042301d379eee6cfbda4d"} Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.018983 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" event={"ID":"738dcd07-f77b-4737-8195-07e674320384","Type":"ContainerDied","Data":"3d56a82b0b92a09fef99b6885f2f82b0534585cd699f41167dee81b33eae30d4"} Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.019024 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d56a82b0b92a09fef99b6885f2f82b0534585cd699f41167dee81b33eae30d4" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.019134 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.025892 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" podStartSLOduration=130.025873741 podStartE2EDuration="2m10.025873741s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:22.022274562 +0000 UTC m=+152.060855732" watchObservedRunningTime="2025-10-07 20:58:22.025873741 +0000 UTC m=+152.064454891" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.324313 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n7ggm"] Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.325526 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.329818 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.333007 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n7ggm"] Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.428217 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.428855 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.432011 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.434960 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.435266 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.449184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-utilities\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.449261 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-catalog-content\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.449312 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h85m8\" (UniqueName: \"kubernetes.io/projected/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-kube-api-access-h85m8\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.550619 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95691413-d284-41a3-90ff-2f2b66ba66d1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.550671 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h85m8\" (UniqueName: \"kubernetes.io/projected/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-kube-api-access-h85m8\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.550762 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-utilities\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.550788 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95691413-d284-41a3-90ff-2f2b66ba66d1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.550816 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-catalog-content\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.551220 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-catalog-content\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.553589 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-utilities\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.584303 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h85m8\" (UniqueName: \"kubernetes.io/projected/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-kube-api-access-h85m8\") pod \"redhat-operators-n7ggm\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.622273 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-w6j8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.622322 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w6j8g" podUID="bf45cac1-ea89-474f-8596-e895d22be327" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.622470 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-w6j8g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.622523 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w6j8g" podUID="bf45cac1-ea89-474f-8596-e895d22be327" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.651719 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.652223 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95691413-d284-41a3-90ff-2f2b66ba66d1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.652298 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95691413-d284-41a3-90ff-2f2b66ba66d1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.652467 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95691413-d284-41a3-90ff-2f2b66ba66d1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.667024 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95691413-d284-41a3-90ff-2f2b66ba66d1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.683417 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:22 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:22 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:22 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.683463 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.722634 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kdpxx"] Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.750048 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kdpxx"] Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.750164 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.753872 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.791422 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.797442 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-n4dfn" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.856031 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-utilities\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.856307 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-catalog-content\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.856342 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzfrv\" (UniqueName: \"kubernetes.io/projected/0c079993-b9ac-4e18-9401-567742e32a0e-kube-api-access-zzfrv\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.864313 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.864888 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.875388 4926 patch_prober.go:28] interesting pod/console-f9d7485db-ggd2q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.875457 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ggd2q" podUID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.959919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-utilities\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.959974 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-catalog-content\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.960014 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzfrv\" (UniqueName: \"kubernetes.io/projected/0c079993-b9ac-4e18-9401-567742e32a0e-kube-api-access-zzfrv\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.961683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-utilities\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.961724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-catalog-content\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:22 crc kubenswrapper[4926]: I1007 20:58:22.987805 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzfrv\" (UniqueName: \"kubernetes.io/projected/0c079993-b9ac-4e18-9401-567742e32a0e-kube-api-access-zzfrv\") pod \"redhat-operators-kdpxx\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.021871 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n7ggm"] Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.032759 4926 generic.go:334] "Generic (PLEG): container finished" podID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerID="9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295" exitCode=0 Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.032840 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7wkx" event={"ID":"7742a5ac-dd42-47b0-af53-077cc6553f78","Type":"ContainerDied","Data":"9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295"} Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.052677 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.066754 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t6nd5" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.070179 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.123606 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 20:58:23 crc kubenswrapper[4926]: W1007 20:58:23.215356 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod95691413_d284_41a3_90ff_2f2b66ba66d1.slice/crio-675c2b644baf4ee652c830279254e456628185601b0357190b81b1869d5efea7 WatchSource:0}: Error finding container 675c2b644baf4ee652c830279254e456628185601b0357190b81b1869d5efea7: Status 404 returned error can't find the container with id 675c2b644baf4ee652c830279254e456628185601b0357190b81b1869d5efea7 Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.319315 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.424919 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kdpxx"] Oct 07 20:58:23 crc kubenswrapper[4926]: W1007 20:58:23.440166 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c079993_b9ac_4e18_9401_567742e32a0e.slice/crio-505d1dfbab90411afc3a4f18037d7e918be7036df73390e0f1b1a2d2787cbb69 WatchSource:0}: Error finding container 505d1dfbab90411afc3a4f18037d7e918be7036df73390e0f1b1a2d2787cbb69: Status 404 returned error can't find the container with id 505d1dfbab90411afc3a4f18037d7e918be7036df73390e0f1b1a2d2787cbb69 Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.487987 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.680402 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.683469 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:23 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:23 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:23 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:23 crc kubenswrapper[4926]: I1007 20:58:23.683516 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.042149 4926 generic.go:334] "Generic (PLEG): container finished" podID="0c079993-b9ac-4e18-9401-567742e32a0e" containerID="8e989de75d6429ff2546cf781055613186d7a4600d8bffb5347b271fae3af818" exitCode=0 Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.042231 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerDied","Data":"8e989de75d6429ff2546cf781055613186d7a4600d8bffb5347b271fae3af818"} Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.044184 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerStarted","Data":"505d1dfbab90411afc3a4f18037d7e918be7036df73390e0f1b1a2d2787cbb69"} Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.049449 4926 generic.go:334] "Generic (PLEG): container finished" podID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerID="c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741" exitCode=0 Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.049543 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7ggm" event={"ID":"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1","Type":"ContainerDied","Data":"c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741"} Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.049574 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7ggm" event={"ID":"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1","Type":"ContainerStarted","Data":"e89e6f9d7583f93d33db7a8c57cdc8a395dafa3493525e562ecd0d167baed435"} Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.055279 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95691413-d284-41a3-90ff-2f2b66ba66d1","Type":"ContainerStarted","Data":"2fe465afc0fb135cca364f72ccf8c463c78b220fa934c52f78cb3f9d5d917354"} Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.055317 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95691413-d284-41a3-90ff-2f2b66ba66d1","Type":"ContainerStarted","Data":"675c2b644baf4ee652c830279254e456628185601b0357190b81b1869d5efea7"} Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.108062 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.108046995 podStartE2EDuration="2.108046995s" podCreationTimestamp="2025-10-07 20:58:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:24.106515889 +0000 UTC m=+154.145097039" watchObservedRunningTime="2025-10-07 20:58:24.108046995 +0000 UTC m=+154.146628135" Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.683313 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:24 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:24 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:24 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:24 crc kubenswrapper[4926]: I1007 20:58:24.683366 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.066638 4926 generic.go:334] "Generic (PLEG): container finished" podID="95691413-d284-41a3-90ff-2f2b66ba66d1" containerID="2fe465afc0fb135cca364f72ccf8c463c78b220fa934c52f78cb3f9d5d917354" exitCode=0 Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.066676 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95691413-d284-41a3-90ff-2f2b66ba66d1","Type":"ContainerDied","Data":"2fe465afc0fb135cca364f72ccf8c463c78b220fa934c52f78cb3f9d5d917354"} Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.171601 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qdvtq" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.475874 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.476665 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.481187 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.483132 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.485722 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.605043 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.605113 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.682646 4926 patch_prober.go:28] interesting pod/router-default-5444994796-7928l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 20:58:25 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Oct 07 20:58:25 crc kubenswrapper[4926]: [+]process-running ok Oct 07 20:58:25 crc kubenswrapper[4926]: healthz check failed Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.682827 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7928l" podUID="1680115c-0bb3-4696-8564-a685fe8efdcb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.706266 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.706383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.706462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.725168 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:25 crc kubenswrapper[4926]: I1007 20:58:25.809733 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:26 crc kubenswrapper[4926]: I1007 20:58:26.045523 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 20:58:26 crc kubenswrapper[4926]: I1007 20:58:26.709431 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:26 crc kubenswrapper[4926]: I1007 20:58:26.724482 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7928l" Oct 07 20:58:31 crc kubenswrapper[4926]: W1007 20:58:31.938697 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod296bedfb_9b6e_4a82_a5ff_480b39b9dff6.slice/crio-5f110112618e97a0a340880b5af9522e9888e5cfecdb86105bae7b819f840e0f WatchSource:0}: Error finding container 5f110112618e97a0a340880b5af9522e9888e5cfecdb86105bae7b819f840e0f: Status 404 returned error can't find the container with id 5f110112618e97a0a340880b5af9522e9888e5cfecdb86105bae7b819f840e0f Oct 07 20:58:31 crc kubenswrapper[4926]: I1007 20:58:31.974181 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.135102 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95691413-d284-41a3-90ff-2f2b66ba66d1","Type":"ContainerDied","Data":"675c2b644baf4ee652c830279254e456628185601b0357190b81b1869d5efea7"} Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.135127 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.135137 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="675c2b644baf4ee652c830279254e456628185601b0357190b81b1869d5efea7" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.136599 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"296bedfb-9b6e-4a82-a5ff-480b39b9dff6","Type":"ContainerStarted","Data":"5f110112618e97a0a340880b5af9522e9888e5cfecdb86105bae7b819f840e0f"} Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.138152 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95691413-d284-41a3-90ff-2f2b66ba66d1-kubelet-dir\") pod \"95691413-d284-41a3-90ff-2f2b66ba66d1\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.138234 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95691413-d284-41a3-90ff-2f2b66ba66d1-kube-api-access\") pod \"95691413-d284-41a3-90ff-2f2b66ba66d1\" (UID: \"95691413-d284-41a3-90ff-2f2b66ba66d1\") " Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.138267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95691413-d284-41a3-90ff-2f2b66ba66d1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "95691413-d284-41a3-90ff-2f2b66ba66d1" (UID: "95691413-d284-41a3-90ff-2f2b66ba66d1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.138534 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95691413-d284-41a3-90ff-2f2b66ba66d1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.146575 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95691413-d284-41a3-90ff-2f2b66ba66d1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "95691413-d284-41a3-90ff-2f2b66ba66d1" (UID: "95691413-d284-41a3-90ff-2f2b66ba66d1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.240275 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95691413-d284-41a3-90ff-2f2b66ba66d1-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.629033 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-w6j8g" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.906135 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:32 crc kubenswrapper[4926]: I1007 20:58:32.910095 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 20:58:33 crc kubenswrapper[4926]: I1007 20:58:33.209404 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 20:58:33 crc kubenswrapper[4926]: I1007 20:58:33.209474 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 20:58:34 crc kubenswrapper[4926]: I1007 20:58:34.480852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:34 crc kubenswrapper[4926]: I1007 20:58:34.492979 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a8e664c-9b14-4510-8a51-bec46ecbaea5-metrics-certs\") pod \"network-metrics-daemon-c6mr4\" (UID: \"7a8e664c-9b14-4510-8a51-bec46ecbaea5\") " pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:34 crc kubenswrapper[4926]: I1007 20:58:34.618655 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c6mr4" Oct 07 20:58:40 crc kubenswrapper[4926]: I1007 20:58:40.471379 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 20:58:44 crc kubenswrapper[4926]: E1007 20:58:44.879979 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 07 20:58:44 crc kubenswrapper[4926]: E1007 20:58:44.880474 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qxvms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wvl5m_openshift-marketplace(fdc058ed-3a5e-4c9c-98ff-6cd72519dd10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 20:58:44 crc kubenswrapper[4926]: E1007 20:58:44.881693 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wvl5m" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" Oct 07 20:58:47 crc kubenswrapper[4926]: E1007 20:58:47.146409 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wvl5m" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.004178 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.004663 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-glmcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-j7wkx_openshift-marketplace(7742a5ac-dd42-47b0-af53-077cc6553f78): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.005724 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-j7wkx" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.019796 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.019933 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-clc9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bpwnc_openshift-marketplace(9035f772-a1f8-425d-90a3-00f038a08bf0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.021004 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bpwnc" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" Oct 07 20:58:49 crc kubenswrapper[4926]: I1007 20:58:49.261045 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerStarted","Data":"e0940d3b0136d7332eb9bd3d7d801d80dc5043b79e63fa969a0a11f8ba188b65"} Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.264216 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-j7wkx" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" Oct 07 20:58:49 crc kubenswrapper[4926]: E1007 20:58:49.264766 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bpwnc" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" Oct 07 20:58:49 crc kubenswrapper[4926]: I1007 20:58:49.340239 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-c6mr4"] Oct 07 20:58:49 crc kubenswrapper[4926]: W1007 20:58:49.348933 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a8e664c_9b14_4510_8a51_bec46ecbaea5.slice/crio-203b67ec893139fe0cc55e99127a21ad54f39b7d981955a2d08c592afe8d1b18 WatchSource:0}: Error finding container 203b67ec893139fe0cc55e99127a21ad54f39b7d981955a2d08c592afe8d1b18: Status 404 returned error can't find the container with id 203b67ec893139fe0cc55e99127a21ad54f39b7d981955a2d08c592afe8d1b18 Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.279690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"296bedfb-9b6e-4a82-a5ff-480b39b9dff6","Type":"ContainerStarted","Data":"10f209515e1baddb1a0da15fd32d1370d443078b887bf6d1be9204ae12edd1de"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.285156 4926 generic.go:334] "Generic (PLEG): container finished" podID="932d000c-28ca-4128-a049-01b7c2960ed1" containerID="9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1" exitCode=0 Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.285263 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hknpx" event={"ID":"932d000c-28ca-4128-a049-01b7c2960ed1","Type":"ContainerDied","Data":"9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.292645 4926 generic.go:334] "Generic (PLEG): container finished" podID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerID="2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a" exitCode=0 Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.292721 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgt9" event={"ID":"86367f52-f7bb-4644-ac82-c1bda5df5784","Type":"ContainerDied","Data":"2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.297485 4926 generic.go:334] "Generic (PLEG): container finished" podID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerID="a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd" exitCode=0 Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.297561 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29lbg" event={"ID":"822a04da-8549-47c1-9994-bcc9b7ca1ed7","Type":"ContainerDied","Data":"a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.302574 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=25.302555388 podStartE2EDuration="25.302555388s" podCreationTimestamp="2025-10-07 20:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:50.295956019 +0000 UTC m=+180.334537199" watchObservedRunningTime="2025-10-07 20:58:50.302555388 +0000 UTC m=+180.341136578" Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.311426 4926 generic.go:334] "Generic (PLEG): container finished" podID="0c079993-b9ac-4e18-9401-567742e32a0e" containerID="e0940d3b0136d7332eb9bd3d7d801d80dc5043b79e63fa969a0a11f8ba188b65" exitCode=0 Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.311520 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerDied","Data":"e0940d3b0136d7332eb9bd3d7d801d80dc5043b79e63fa969a0a11f8ba188b65"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.314560 4926 generic.go:334] "Generic (PLEG): container finished" podID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerID="0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5" exitCode=0 Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.314628 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7ggm" event={"ID":"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1","Type":"ContainerDied","Data":"0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.323341 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" event={"ID":"7a8e664c-9b14-4510-8a51-bec46ecbaea5","Type":"ContainerStarted","Data":"3a66dc57b457776536b36b67685d54057f54493383ec34afd966e2bbfd61cc89"} Oct 07 20:58:50 crc kubenswrapper[4926]: I1007 20:58:50.323388 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" event={"ID":"7a8e664c-9b14-4510-8a51-bec46ecbaea5","Type":"ContainerStarted","Data":"203b67ec893139fe0cc55e99127a21ad54f39b7d981955a2d08c592afe8d1b18"} Oct 07 20:58:51 crc kubenswrapper[4926]: I1007 20:58:51.331745 4926 generic.go:334] "Generic (PLEG): container finished" podID="296bedfb-9b6e-4a82-a5ff-480b39b9dff6" containerID="10f209515e1baddb1a0da15fd32d1370d443078b887bf6d1be9204ae12edd1de" exitCode=0 Oct 07 20:58:51 crc kubenswrapper[4926]: I1007 20:58:51.331928 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"296bedfb-9b6e-4a82-a5ff-480b39b9dff6","Type":"ContainerDied","Data":"10f209515e1baddb1a0da15fd32d1370d443078b887bf6d1be9204ae12edd1de"} Oct 07 20:58:51 crc kubenswrapper[4926]: I1007 20:58:51.334490 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c6mr4" event={"ID":"7a8e664c-9b14-4510-8a51-bec46ecbaea5","Type":"ContainerStarted","Data":"08c036672f428c93ae9b732ac00248bf88380b5b5547243c45e002addc6f2342"} Oct 07 20:58:51 crc kubenswrapper[4926]: I1007 20:58:51.374011 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-c6mr4" podStartSLOduration=159.373988311 podStartE2EDuration="2m39.373988311s" podCreationTimestamp="2025-10-07 20:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:58:51.370421353 +0000 UTC m=+181.409002503" watchObservedRunningTime="2025-10-07 20:58:51.373988311 +0000 UTC m=+181.412569461" Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.748739 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.878195 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kube-api-access\") pod \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.878434 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kubelet-dir\") pod \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\" (UID: \"296bedfb-9b6e-4a82-a5ff-480b39b9dff6\") " Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.878525 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "296bedfb-9b6e-4a82-a5ff-480b39b9dff6" (UID: "296bedfb-9b6e-4a82-a5ff-480b39b9dff6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.879818 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.884185 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "296bedfb-9b6e-4a82-a5ff-480b39b9dff6" (UID: "296bedfb-9b6e-4a82-a5ff-480b39b9dff6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:58:52 crc kubenswrapper[4926]: I1007 20:58:52.980843 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/296bedfb-9b6e-4a82-a5ff-480b39b9dff6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 20:58:53 crc kubenswrapper[4926]: I1007 20:58:53.355778 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hknpx" event={"ID":"932d000c-28ca-4128-a049-01b7c2960ed1","Type":"ContainerStarted","Data":"2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a"} Oct 07 20:58:53 crc kubenswrapper[4926]: I1007 20:58:53.357968 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"296bedfb-9b6e-4a82-a5ff-480b39b9dff6","Type":"ContainerDied","Data":"5f110112618e97a0a340880b5af9522e9888e5cfecdb86105bae7b819f840e0f"} Oct 07 20:58:53 crc kubenswrapper[4926]: I1007 20:58:53.358024 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f110112618e97a0a340880b5af9522e9888e5cfecdb86105bae7b819f840e0f" Oct 07 20:58:53 crc kubenswrapper[4926]: I1007 20:58:53.358078 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 20:58:53 crc kubenswrapper[4926]: I1007 20:58:53.643372 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-57mj8" Oct 07 20:58:54 crc kubenswrapper[4926]: I1007 20:58:54.390082 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hknpx" podStartSLOduration=3.528755417 podStartE2EDuration="35.390056227s" podCreationTimestamp="2025-10-07 20:58:19 +0000 UTC" firstStartedPulling="2025-10-07 20:58:20.955211071 +0000 UTC m=+150.993792221" lastFinishedPulling="2025-10-07 20:58:52.816511881 +0000 UTC m=+182.855093031" observedRunningTime="2025-10-07 20:58:54.384552051 +0000 UTC m=+184.423133241" watchObservedRunningTime="2025-10-07 20:58:54.390056227 +0000 UTC m=+184.428637417" Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.371052 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29lbg" event={"ID":"822a04da-8549-47c1-9994-bcc9b7ca1ed7","Type":"ContainerStarted","Data":"8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2"} Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.373797 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerStarted","Data":"fc00a0cb45d18f4a7245cef256376942dae1f4aaf27a30ff444c1269aed11c12"} Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.375491 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7ggm" event={"ID":"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1","Type":"ContainerStarted","Data":"ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d"} Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.378241 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgt9" event={"ID":"86367f52-f7bb-4644-ac82-c1bda5df5784","Type":"ContainerStarted","Data":"72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4"} Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.409898 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kdpxx" podStartSLOduration=2.756536848 podStartE2EDuration="33.409885892s" podCreationTimestamp="2025-10-07 20:58:22 +0000 UTC" firstStartedPulling="2025-10-07 20:58:24.046625301 +0000 UTC m=+154.085206451" lastFinishedPulling="2025-10-07 20:58:54.699974345 +0000 UTC m=+184.738555495" observedRunningTime="2025-10-07 20:58:55.409443669 +0000 UTC m=+185.448024829" watchObservedRunningTime="2025-10-07 20:58:55.409885892 +0000 UTC m=+185.448467042" Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.411347 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-29lbg" podStartSLOduration=3.605046293 podStartE2EDuration="37.411341486s" podCreationTimestamp="2025-10-07 20:58:18 +0000 UTC" firstStartedPulling="2025-10-07 20:58:20.9697803 +0000 UTC m=+151.008361450" lastFinishedPulling="2025-10-07 20:58:54.776075483 +0000 UTC m=+184.814656643" observedRunningTime="2025-10-07 20:58:55.39357634 +0000 UTC m=+185.432157490" watchObservedRunningTime="2025-10-07 20:58:55.411341486 +0000 UTC m=+185.449922636" Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.425135 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbgt9" podStartSLOduration=3.18782492 podStartE2EDuration="36.425127042s" podCreationTimestamp="2025-10-07 20:58:19 +0000 UTC" firstStartedPulling="2025-10-07 20:58:20.962591143 +0000 UTC m=+151.001172293" lastFinishedPulling="2025-10-07 20:58:54.199893265 +0000 UTC m=+184.238474415" observedRunningTime="2025-10-07 20:58:55.423653908 +0000 UTC m=+185.462235058" watchObservedRunningTime="2025-10-07 20:58:55.425127042 +0000 UTC m=+185.463708192" Oct 07 20:58:55 crc kubenswrapper[4926]: I1007 20:58:55.443197 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n7ggm" podStartSLOduration=2.659438555 podStartE2EDuration="33.443180327s" podCreationTimestamp="2025-10-07 20:58:22 +0000 UTC" firstStartedPulling="2025-10-07 20:58:24.053164578 +0000 UTC m=+154.091745728" lastFinishedPulling="2025-10-07 20:58:54.83690634 +0000 UTC m=+184.875487500" observedRunningTime="2025-10-07 20:58:55.439825016 +0000 UTC m=+185.478406166" watchObservedRunningTime="2025-10-07 20:58:55.443180327 +0000 UTC m=+185.481761477" Oct 07 20:58:58 crc kubenswrapper[4926]: I1007 20:58:58.815249 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.338019 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.338265 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.513853 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.560909 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.560960 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.610638 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.845708 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.846051 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:58:59 crc kubenswrapper[4926]: I1007 20:58:59.885190 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:59:00 crc kubenswrapper[4926]: I1007 20:59:00.438774 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 20:59:00 crc kubenswrapper[4926]: I1007 20:59:00.457307 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:59:00 crc kubenswrapper[4926]: I1007 20:59:00.481415 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-29lbg" Oct 07 20:59:01 crc kubenswrapper[4926]: I1007 20:59:01.912403 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hknpx"] Oct 07 20:59:02 crc kubenswrapper[4926]: I1007 20:59:02.413561 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hknpx" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="registry-server" containerID="cri-o://2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a" gracePeriod=2 Oct 07 20:59:02 crc kubenswrapper[4926]: I1007 20:59:02.652335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:59:02 crc kubenswrapper[4926]: I1007 20:59:02.652978 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:59:02 crc kubenswrapper[4926]: I1007 20:59:02.702298 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.070752 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.070800 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.123476 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.213430 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.213501 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.344898 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.436356 4926 generic.go:334] "Generic (PLEG): container finished" podID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerID="821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1" exitCode=0 Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.436623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpwnc" event={"ID":"9035f772-a1f8-425d-90a3-00f038a08bf0","Type":"ContainerDied","Data":"821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1"} Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.440812 4926 generic.go:334] "Generic (PLEG): container finished" podID="932d000c-28ca-4128-a049-01b7c2960ed1" containerID="2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a" exitCode=0 Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.440879 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hknpx" event={"ID":"932d000c-28ca-4128-a049-01b7c2960ed1","Type":"ContainerDied","Data":"2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a"} Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.440905 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hknpx" event={"ID":"932d000c-28ca-4128-a049-01b7c2960ed1","Type":"ContainerDied","Data":"586f609eb95df2e81f9cbdb952018e729d48a0cb39cd600ff2b8f2746a9d840d"} Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.440922 4926 scope.go:117] "RemoveContainer" containerID="2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.441182 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hknpx" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.450113 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerStarted","Data":"b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407"} Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.461031 4926 scope.go:117] "RemoveContainer" containerID="9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.484608 4926 scope.go:117] "RemoveContainer" containerID="f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.497613 4926 scope.go:117] "RemoveContainer" containerID="2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.500043 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 20:59:03 crc kubenswrapper[4926]: E1007 20:59:03.501695 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a\": container with ID starting with 2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a not found: ID does not exist" containerID="2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.501782 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a"} err="failed to get container status \"2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a\": rpc error: code = NotFound desc = could not find container \"2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a\": container with ID starting with 2ad8cb1a77f522ab599e8f4c1cb9f7bedf761ca6c277516b316e7c24c39c838a not found: ID does not exist" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.501865 4926 scope.go:117] "RemoveContainer" containerID="9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1" Oct 07 20:59:03 crc kubenswrapper[4926]: E1007 20:59:03.502729 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1\": container with ID starting with 9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1 not found: ID does not exist" containerID="9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.502798 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1"} err="failed to get container status \"9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1\": rpc error: code = NotFound desc = could not find container \"9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1\": container with ID starting with 9120c752bbd89e639263772d8da216ab17a5bce307500be773aa0da88e86dfb1 not found: ID does not exist" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.502849 4926 scope.go:117] "RemoveContainer" containerID="f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074" Oct 07 20:59:03 crc kubenswrapper[4926]: E1007 20:59:03.503317 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074\": container with ID starting with f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074 not found: ID does not exist" containerID="f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.503369 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074"} err="failed to get container status \"f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074\": rpc error: code = NotFound desc = could not find container \"f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074\": container with ID starting with f9d1706f91debaa7bc58a04001091e015bd450301ac032ad538a4b34ab579074 not found: ID does not exist" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.504329 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.531091 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-catalog-content\") pod \"932d000c-28ca-4128-a049-01b7c2960ed1\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.531210 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-utilities\") pod \"932d000c-28ca-4128-a049-01b7c2960ed1\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.531247 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rjh8\" (UniqueName: \"kubernetes.io/projected/932d000c-28ca-4128-a049-01b7c2960ed1-kube-api-access-4rjh8\") pod \"932d000c-28ca-4128-a049-01b7c2960ed1\" (UID: \"932d000c-28ca-4128-a049-01b7c2960ed1\") " Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.532094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-utilities" (OuterVolumeSpecName: "utilities") pod "932d000c-28ca-4128-a049-01b7c2960ed1" (UID: "932d000c-28ca-4128-a049-01b7c2960ed1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.567115 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/932d000c-28ca-4128-a049-01b7c2960ed1-kube-api-access-4rjh8" (OuterVolumeSpecName: "kube-api-access-4rjh8") pod "932d000c-28ca-4128-a049-01b7c2960ed1" (UID: "932d000c-28ca-4128-a049-01b7c2960ed1"). InnerVolumeSpecName "kube-api-access-4rjh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.582681 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "932d000c-28ca-4128-a049-01b7c2960ed1" (UID: "932d000c-28ca-4128-a049-01b7c2960ed1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.633405 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.633437 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/932d000c-28ca-4128-a049-01b7c2960ed1-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.633449 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rjh8\" (UniqueName: \"kubernetes.io/projected/932d000c-28ca-4128-a049-01b7c2960ed1-kube-api-access-4rjh8\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.767998 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hknpx"] Oct 07 20:59:03 crc kubenswrapper[4926]: I1007 20:59:03.772838 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hknpx"] Oct 07 20:59:04 crc kubenswrapper[4926]: I1007 20:59:04.457629 4926 generic.go:334] "Generic (PLEG): container finished" podID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerID="b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407" exitCode=0 Oct 07 20:59:04 crc kubenswrapper[4926]: I1007 20:59:04.457712 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerDied","Data":"b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407"} Oct 07 20:59:04 crc kubenswrapper[4926]: I1007 20:59:04.461699 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpwnc" event={"ID":"9035f772-a1f8-425d-90a3-00f038a08bf0","Type":"ContainerStarted","Data":"d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef"} Oct 07 20:59:04 crc kubenswrapper[4926]: I1007 20:59:04.499180 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bpwnc" podStartSLOduration=2.666685872 podStartE2EDuration="44.499155907s" podCreationTimestamp="2025-10-07 20:58:20 +0000 UTC" firstStartedPulling="2025-10-07 20:58:22.017150338 +0000 UTC m=+152.055731488" lastFinishedPulling="2025-10-07 20:59:03.849620363 +0000 UTC m=+193.888201523" observedRunningTime="2025-10-07 20:59:04.495856737 +0000 UTC m=+194.534437887" watchObservedRunningTime="2025-10-07 20:59:04.499155907 +0000 UTC m=+194.537737047" Oct 07 20:59:04 crc kubenswrapper[4926]: I1007 20:59:04.687163 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" path="/var/lib/kubelet/pods/932d000c-28ca-4128-a049-01b7c2960ed1/volumes" Oct 07 20:59:05 crc kubenswrapper[4926]: I1007 20:59:05.474769 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerStarted","Data":"13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959"} Oct 07 20:59:05 crc kubenswrapper[4926]: I1007 20:59:05.476943 4926 generic.go:334] "Generic (PLEG): container finished" podID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerID="f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa" exitCode=0 Oct 07 20:59:05 crc kubenswrapper[4926]: I1007 20:59:05.477003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7wkx" event={"ID":"7742a5ac-dd42-47b0-af53-077cc6553f78","Type":"ContainerDied","Data":"f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa"} Oct 07 20:59:05 crc kubenswrapper[4926]: I1007 20:59:05.495950 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wvl5m" podStartSLOduration=2.468649154 podStartE2EDuration="46.495927915s" podCreationTimestamp="2025-10-07 20:58:19 +0000 UTC" firstStartedPulling="2025-10-07 20:58:20.937914948 +0000 UTC m=+150.976496098" lastFinishedPulling="2025-10-07 20:59:04.965193719 +0000 UTC m=+195.003774859" observedRunningTime="2025-10-07 20:59:05.49343417 +0000 UTC m=+195.532015320" watchObservedRunningTime="2025-10-07 20:59:05.495927915 +0000 UTC m=+195.534509065" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.306111 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kdpxx"] Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.306751 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kdpxx" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="registry-server" containerID="cri-o://fc00a0cb45d18f4a7245cef256376942dae1f4aaf27a30ff444c1269aed11c12" gracePeriod=2 Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.484575 4926 generic.go:334] "Generic (PLEG): container finished" podID="0c079993-b9ac-4e18-9401-567742e32a0e" containerID="fc00a0cb45d18f4a7245cef256376942dae1f4aaf27a30ff444c1269aed11c12" exitCode=0 Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.484635 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerDied","Data":"fc00a0cb45d18f4a7245cef256376942dae1f4aaf27a30ff444c1269aed11c12"} Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.488244 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7wkx" event={"ID":"7742a5ac-dd42-47b0-af53-077cc6553f78","Type":"ContainerStarted","Data":"628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e"} Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.502391 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j7wkx" podStartSLOduration=2.32936067 podStartE2EDuration="45.502376456s" podCreationTimestamp="2025-10-07 20:58:21 +0000 UTC" firstStartedPulling="2025-10-07 20:58:23.038802597 +0000 UTC m=+153.077383747" lastFinishedPulling="2025-10-07 20:59:06.211818373 +0000 UTC m=+196.250399533" observedRunningTime="2025-10-07 20:59:06.502037456 +0000 UTC m=+196.540618606" watchObservedRunningTime="2025-10-07 20:59:06.502376456 +0000 UTC m=+196.540957606" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.646030 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.783144 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-utilities\") pod \"0c079993-b9ac-4e18-9401-567742e32a0e\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.783491 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-catalog-content\") pod \"0c079993-b9ac-4e18-9401-567742e32a0e\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.783524 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzfrv\" (UniqueName: \"kubernetes.io/projected/0c079993-b9ac-4e18-9401-567742e32a0e-kube-api-access-zzfrv\") pod \"0c079993-b9ac-4e18-9401-567742e32a0e\" (UID: \"0c079993-b9ac-4e18-9401-567742e32a0e\") " Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.784167 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-utilities" (OuterVolumeSpecName: "utilities") pod "0c079993-b9ac-4e18-9401-567742e32a0e" (UID: "0c079993-b9ac-4e18-9401-567742e32a0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.788414 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c079993-b9ac-4e18-9401-567742e32a0e-kube-api-access-zzfrv" (OuterVolumeSpecName: "kube-api-access-zzfrv") pod "0c079993-b9ac-4e18-9401-567742e32a0e" (UID: "0c079993-b9ac-4e18-9401-567742e32a0e"). InnerVolumeSpecName "kube-api-access-zzfrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.793795 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.793819 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzfrv\" (UniqueName: \"kubernetes.io/projected/0c079993-b9ac-4e18-9401-567742e32a0e-kube-api-access-zzfrv\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.868663 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c079993-b9ac-4e18-9401-567742e32a0e" (UID: "0c079993-b9ac-4e18-9401-567742e32a0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:06 crc kubenswrapper[4926]: I1007 20:59:06.894353 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c079993-b9ac-4e18-9401-567742e32a0e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.503090 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdpxx" event={"ID":"0c079993-b9ac-4e18-9401-567742e32a0e","Type":"ContainerDied","Data":"505d1dfbab90411afc3a4f18037d7e918be7036df73390e0f1b1a2d2787cbb69"} Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.503138 4926 scope.go:117] "RemoveContainer" containerID="fc00a0cb45d18f4a7245cef256376942dae1f4aaf27a30ff444c1269aed11c12" Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.503152 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdpxx" Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.520341 4926 scope.go:117] "RemoveContainer" containerID="e0940d3b0136d7332eb9bd3d7d801d80dc5043b79e63fa969a0a11f8ba188b65" Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.528026 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kdpxx"] Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.532156 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kdpxx"] Oct 07 20:59:07 crc kubenswrapper[4926]: I1007 20:59:07.544524 4926 scope.go:117] "RemoveContainer" containerID="8e989de75d6429ff2546cf781055613186d7a4600d8bffb5347b271fae3af818" Oct 07 20:59:08 crc kubenswrapper[4926]: I1007 20:59:08.684382 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" path="/var/lib/kubelet/pods/0c079993-b9ac-4e18-9401-567742e32a0e/volumes" Oct 07 20:59:09 crc kubenswrapper[4926]: I1007 20:59:09.750127 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:59:09 crc kubenswrapper[4926]: I1007 20:59:09.750175 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:59:09 crc kubenswrapper[4926]: I1007 20:59:09.783107 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:59:10 crc kubenswrapper[4926]: I1007 20:59:10.576487 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.300950 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.301338 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.309942 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wvl5m"] Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.344151 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.560026 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.670238 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.670280 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:59:11 crc kubenswrapper[4926]: I1007 20:59:11.720607 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.527685 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wvl5m" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="registry-server" containerID="cri-o://13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959" gracePeriod=2 Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.586975 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.889120 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.963364 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-catalog-content\") pod \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.963439 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxvms\" (UniqueName: \"kubernetes.io/projected/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-kube-api-access-qxvms\") pod \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.963554 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-utilities\") pod \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\" (UID: \"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10\") " Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.964765 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-utilities" (OuterVolumeSpecName: "utilities") pod "fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" (UID: "fdc058ed-3a5e-4c9c-98ff-6cd72519dd10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.964887 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:12 crc kubenswrapper[4926]: I1007 20:59:12.968631 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-kube-api-access-qxvms" (OuterVolumeSpecName: "kube-api-access-qxvms") pod "fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" (UID: "fdc058ed-3a5e-4c9c-98ff-6cd72519dd10"). InnerVolumeSpecName "kube-api-access-qxvms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.017330 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" (UID: "fdc058ed-3a5e-4c9c-98ff-6cd72519dd10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.065538 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.065573 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxvms\" (UniqueName: \"kubernetes.io/projected/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10-kube-api-access-qxvms\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.534709 4926 generic.go:334] "Generic (PLEG): container finished" podID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerID="13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959" exitCode=0 Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.534787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerDied","Data":"13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959"} Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.535118 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvl5m" event={"ID":"fdc058ed-3a5e-4c9c-98ff-6cd72519dd10","Type":"ContainerDied","Data":"8c624bd212a789b1db559307e3adcf1eb5cf3e5046d7827f4e62475c423ba647"} Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.534834 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvl5m" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.535161 4926 scope.go:117] "RemoveContainer" containerID="13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.549932 4926 scope.go:117] "RemoveContainer" containerID="b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.567038 4926 scope.go:117] "RemoveContainer" containerID="dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.571158 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wvl5m"] Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.580019 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wvl5m"] Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.591823 4926 scope.go:117] "RemoveContainer" containerID="13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959" Oct 07 20:59:13 crc kubenswrapper[4926]: E1007 20:59:13.596369 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959\": container with ID starting with 13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959 not found: ID does not exist" containerID="13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.596443 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959"} err="failed to get container status \"13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959\": rpc error: code = NotFound desc = could not find container \"13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959\": container with ID starting with 13e0186e5303bba9a9de42392a6a36cffb555e4e2e9a0d7a5606e5161cd39959 not found: ID does not exist" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.596484 4926 scope.go:117] "RemoveContainer" containerID="b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407" Oct 07 20:59:13 crc kubenswrapper[4926]: E1007 20:59:13.596866 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407\": container with ID starting with b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407 not found: ID does not exist" containerID="b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.596914 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407"} err="failed to get container status \"b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407\": rpc error: code = NotFound desc = could not find container \"b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407\": container with ID starting with b79f59f31bb81c45d1f6d71f3f187655e0c689bea49527d7b726c22b7b47b407 not found: ID does not exist" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.596946 4926 scope.go:117] "RemoveContainer" containerID="dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f" Oct 07 20:59:13 crc kubenswrapper[4926]: E1007 20:59:13.597412 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f\": container with ID starting with dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f not found: ID does not exist" containerID="dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f" Oct 07 20:59:13 crc kubenswrapper[4926]: I1007 20:59:13.597459 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f"} err="failed to get container status \"dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f\": rpc error: code = NotFound desc = could not find container \"dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f\": container with ID starting with dd8aa511f22696ad2b195736a709996f75dab6f34bd7565e5dd8bb35d00db36f not found: ID does not exist" Oct 07 20:59:14 crc kubenswrapper[4926]: I1007 20:59:14.686047 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" path="/var/lib/kubelet/pods/fdc058ed-3a5e-4c9c-98ff-6cd72519dd10/volumes" Oct 07 20:59:16 crc kubenswrapper[4926]: I1007 20:59:16.720813 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7wkx"] Oct 07 20:59:16 crc kubenswrapper[4926]: I1007 20:59:16.721320 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j7wkx" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="registry-server" containerID="cri-o://628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e" gracePeriod=2 Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.096257 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.226190 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-catalog-content\") pod \"7742a5ac-dd42-47b0-af53-077cc6553f78\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.226315 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-utilities\") pod \"7742a5ac-dd42-47b0-af53-077cc6553f78\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.226428 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glmcp\" (UniqueName: \"kubernetes.io/projected/7742a5ac-dd42-47b0-af53-077cc6553f78-kube-api-access-glmcp\") pod \"7742a5ac-dd42-47b0-af53-077cc6553f78\" (UID: \"7742a5ac-dd42-47b0-af53-077cc6553f78\") " Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.226978 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-utilities" (OuterVolumeSpecName: "utilities") pod "7742a5ac-dd42-47b0-af53-077cc6553f78" (UID: "7742a5ac-dd42-47b0-af53-077cc6553f78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.235639 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7742a5ac-dd42-47b0-af53-077cc6553f78-kube-api-access-glmcp" (OuterVolumeSpecName: "kube-api-access-glmcp") pod "7742a5ac-dd42-47b0-af53-077cc6553f78" (UID: "7742a5ac-dd42-47b0-af53-077cc6553f78"). InnerVolumeSpecName "kube-api-access-glmcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.245142 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7742a5ac-dd42-47b0-af53-077cc6553f78" (UID: "7742a5ac-dd42-47b0-af53-077cc6553f78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.328062 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glmcp\" (UniqueName: \"kubernetes.io/projected/7742a5ac-dd42-47b0-af53-077cc6553f78-kube-api-access-glmcp\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.328097 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.328110 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7742a5ac-dd42-47b0-af53-077cc6553f78-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.564125 4926 generic.go:334] "Generic (PLEG): container finished" podID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerID="628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e" exitCode=0 Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.564180 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7wkx" event={"ID":"7742a5ac-dd42-47b0-af53-077cc6553f78","Type":"ContainerDied","Data":"628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e"} Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.564237 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7wkx" event={"ID":"7742a5ac-dd42-47b0-af53-077cc6553f78","Type":"ContainerDied","Data":"a0cda9d2a91bb86cc4be854118dc1c7a318d11e5e9756e744b9e71f8e47fa39e"} Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.564263 4926 scope.go:117] "RemoveContainer" containerID="628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.564426 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7wkx" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.582664 4926 scope.go:117] "RemoveContainer" containerID="f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.599436 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7wkx"] Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.602018 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7wkx"] Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.603885 4926 scope.go:117] "RemoveContainer" containerID="9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.633542 4926 scope.go:117] "RemoveContainer" containerID="628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e" Oct 07 20:59:17 crc kubenswrapper[4926]: E1007 20:59:17.634306 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e\": container with ID starting with 628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e not found: ID does not exist" containerID="628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.634342 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e"} err="failed to get container status \"628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e\": rpc error: code = NotFound desc = could not find container \"628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e\": container with ID starting with 628e0fdc9b7ae771660cd3a6160e733509f668404306b84ae99b5080a42b672e not found: ID does not exist" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.634370 4926 scope.go:117] "RemoveContainer" containerID="f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa" Oct 07 20:59:17 crc kubenswrapper[4926]: E1007 20:59:17.634761 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa\": container with ID starting with f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa not found: ID does not exist" containerID="f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.634792 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa"} err="failed to get container status \"f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa\": rpc error: code = NotFound desc = could not find container \"f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa\": container with ID starting with f3628165420a6324c9432e248b5ed458ff1718a2b6a69bc72422c16313305caa not found: ID does not exist" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.634856 4926 scope.go:117] "RemoveContainer" containerID="9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295" Oct 07 20:59:17 crc kubenswrapper[4926]: E1007 20:59:17.635222 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295\": container with ID starting with 9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295 not found: ID does not exist" containerID="9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295" Oct 07 20:59:17 crc kubenswrapper[4926]: I1007 20:59:17.635248 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295"} err="failed to get container status \"9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295\": rpc error: code = NotFound desc = could not find container \"9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295\": container with ID starting with 9fcdfbf4023eaad7cedc998e7653a31fea11785b6ed8e33e74752c6722238295 not found: ID does not exist" Oct 07 20:59:18 crc kubenswrapper[4926]: I1007 20:59:18.687010 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" path="/var/lib/kubelet/pods/7742a5ac-dd42-47b0-af53-077cc6553f78/volumes" Oct 07 20:59:21 crc kubenswrapper[4926]: I1007 20:59:21.943344 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ml4zz"] Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.209836 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.210531 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.210595 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.211375 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.211469 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327" gracePeriod=600 Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.672271 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327" exitCode=0 Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.672400 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327"} Oct 07 20:59:33 crc kubenswrapper[4926]: I1007 20:59:33.672697 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"03c2adc15ef42fc76b9b086cb5172f247b75e21cd7f7824690f65b2c4a8e8842"} Oct 07 20:59:46 crc kubenswrapper[4926]: I1007 20:59:46.983110 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" podUID="1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" containerName="oauth-openshift" containerID="cri-o://8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741" gracePeriod=15 Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.387845 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436117 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-578876cb6-gfx9f"] Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436435 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95691413-d284-41a3-90ff-2f2b66ba66d1" containerName="pruner" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436460 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="95691413-d284-41a3-90ff-2f2b66ba66d1" containerName="pruner" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436479 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436492 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436512 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436528 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436544 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436556 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436572 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436584 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436602 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436614 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436628 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296bedfb-9b6e-4a82-a5ff-480b39b9dff6" containerName="pruner" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436639 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="296bedfb-9b6e-4a82-a5ff-480b39b9dff6" containerName="pruner" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436661 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436674 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436689 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436702 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436719 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436730 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436747 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436759 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="extract-utilities" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436773 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436784 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436801 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" containerName="oauth-openshift" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436812 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" containerName="oauth-openshift" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436827 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436838 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.436855 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.436867 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="extract-content" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437018 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="296bedfb-9b6e-4a82-a5ff-480b39b9dff6" containerName="pruner" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437038 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="932d000c-28ca-4128-a049-01b7c2960ed1" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437052 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7742a5ac-dd42-47b0-af53-077cc6553f78" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437069 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" containerName="oauth-openshift" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437087 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c079993-b9ac-4e18-9401-567742e32a0e" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437106 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="95691413-d284-41a3-90ff-2f2b66ba66d1" containerName="pruner" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437122 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdc058ed-3a5e-4c9c-98ff-6cd72519dd10" containerName="registry-server" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.437761 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.464996 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-578876cb6-gfx9f"] Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575651 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-trusted-ca-bundle\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575751 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-error\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575807 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-idp-0-file-data\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575860 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-cliconfig\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575899 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-login\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575929 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-dir\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.575981 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-ocp-branding-template\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576030 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-provider-selection\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576065 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-router-certs\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576060 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576104 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28ccl\" (UniqueName: \"kubernetes.io/projected/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-kube-api-access-28ccl\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576145 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-session\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576224 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-service-ca\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576306 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-policies\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576352 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-serving-cert\") pod \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\" (UID: \"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11\") " Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576552 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f17d30c5-baea-46d9-b1b8-7368aab6480b-audit-dir\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576605 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-session\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576637 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576652 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576691 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-audit-policies\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576725 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-service-ca\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576762 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576803 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576843 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-router-certs\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576880 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/f17d30c5-baea-46d9-b1b8-7368aab6480b-kube-api-access-tdt4c\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.576715 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577061 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577092 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-login\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577227 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-error\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577271 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577337 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577391 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577407 4926 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577423 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577434 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.577732 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.581392 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.581828 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.582883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.583347 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.583728 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.583499 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.584707 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-kube-api-access-28ccl" (OuterVolumeSpecName: "kube-api-access-28ccl") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "kube-api-access-28ccl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.587534 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.593578 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" (UID: "1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678446 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678545 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678599 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f17d30c5-baea-46d9-b1b8-7368aab6480b-audit-dir\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678643 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-session\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678725 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-audit-policies\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678759 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-service-ca\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678760 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f17d30c5-baea-46d9-b1b8-7368aab6480b-audit-dir\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.678903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.679047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-router-certs\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.679359 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/f17d30c5-baea-46d9-b1b8-7368aab6480b-kube-api-access-tdt4c\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.679522 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-login\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.679608 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-error\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680011 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680090 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680110 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680129 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680145 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680163 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680183 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680345 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680366 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680396 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680413 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28ccl\" (UniqueName: \"kubernetes.io/projected/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11-kube-api-access-28ccl\") on node \"crc\" DevicePath \"\"" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680843 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-service-ca\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.680886 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-audit-policies\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.681337 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.682486 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.682764 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.684241 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-session\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.684483 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.684727 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.687823 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-error\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.687976 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.688187 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-system-router-certs\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.688381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f17d30c5-baea-46d9-b1b8-7368aab6480b-v4-0-config-user-template-login\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.704657 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/f17d30c5-baea-46d9-b1b8-7368aab6480b-kube-api-access-tdt4c\") pod \"oauth-openshift-578876cb6-gfx9f\" (UID: \"f17d30c5-baea-46d9-b1b8-7368aab6480b\") " pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.749779 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.786110 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" containerID="8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741" exitCode=0 Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.786241 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.786277 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" event={"ID":"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11","Type":"ContainerDied","Data":"8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741"} Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.787323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ml4zz" event={"ID":"1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11","Type":"ContainerDied","Data":"8092dca64f8ef7c434d5afc6dbd7db51b2125a67159219c169d5cc3f6561547a"} Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.787419 4926 scope.go:117] "RemoveContainer" containerID="8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.820743 4926 scope.go:117] "RemoveContainer" containerID="8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741" Oct 07 20:59:47 crc kubenswrapper[4926]: E1007 20:59:47.821782 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741\": container with ID starting with 8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741 not found: ID does not exist" containerID="8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.821851 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741"} err="failed to get container status \"8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741\": rpc error: code = NotFound desc = could not find container \"8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741\": container with ID starting with 8f024e3b73a5111f35eff66de3a0e9f19da0a514f8c2e35e9237202124c69741 not found: ID does not exist" Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.833573 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ml4zz"] Oct 07 20:59:47 crc kubenswrapper[4926]: I1007 20:59:47.836486 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ml4zz"] Oct 07 20:59:48 crc kubenswrapper[4926]: I1007 20:59:48.231118 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-578876cb6-gfx9f"] Oct 07 20:59:48 crc kubenswrapper[4926]: I1007 20:59:48.699814 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11" path="/var/lib/kubelet/pods/1ed9e5c4-e708-42b2-ac9c-0bfc7a17cf11/volumes" Oct 07 20:59:48 crc kubenswrapper[4926]: I1007 20:59:48.795608 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" event={"ID":"f17d30c5-baea-46d9-b1b8-7368aab6480b","Type":"ContainerStarted","Data":"808611faf9ea8a103dccdd9bac8cc5f9ee8786988cd1c633dead0e5640f47868"} Oct 07 20:59:48 crc kubenswrapper[4926]: I1007 20:59:48.796289 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:48 crc kubenswrapper[4926]: I1007 20:59:48.796329 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" event={"ID":"f17d30c5-baea-46d9-b1b8-7368aab6480b","Type":"ContainerStarted","Data":"6bec07a22a455e0bf46d1056c95d10940a8f13dab5fc124367c38bfc23ca5341"} Oct 07 20:59:49 crc kubenswrapper[4926]: I1007 20:59:49.091391 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" Oct 07 20:59:49 crc kubenswrapper[4926]: I1007 20:59:49.125924 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-578876cb6-gfx9f" podStartSLOduration=28.125897983 podStartE2EDuration="28.125897983s" podCreationTimestamp="2025-10-07 20:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 20:59:48.832686264 +0000 UTC m=+238.871267474" watchObservedRunningTime="2025-10-07 20:59:49.125897983 +0000 UTC m=+239.164479163" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.135397 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v"] Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.167525 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v"] Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.167708 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.169542 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.170039 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.271459 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/43588f58-5468-4b85-8678-58bdd40823c9-config-volume\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.271556 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/43588f58-5468-4b85-8678-58bdd40823c9-secret-volume\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.271587 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwn8x\" (UniqueName: \"kubernetes.io/projected/43588f58-5468-4b85-8678-58bdd40823c9-kube-api-access-wwn8x\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.373184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/43588f58-5468-4b85-8678-58bdd40823c9-secret-volume\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.373264 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwn8x\" (UniqueName: \"kubernetes.io/projected/43588f58-5468-4b85-8678-58bdd40823c9-kube-api-access-wwn8x\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.373365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/43588f58-5468-4b85-8678-58bdd40823c9-config-volume\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.374709 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/43588f58-5468-4b85-8678-58bdd40823c9-config-volume\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.383372 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/43588f58-5468-4b85-8678-58bdd40823c9-secret-volume\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.406281 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwn8x\" (UniqueName: \"kubernetes.io/projected/43588f58-5468-4b85-8678-58bdd40823c9-kube-api-access-wwn8x\") pod \"collect-profiles-29331180-48f9v\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.493467 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:00 crc kubenswrapper[4926]: I1007 21:00:00.921351 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v"] Oct 07 21:00:01 crc kubenswrapper[4926]: I1007 21:00:01.882433 4926 generic.go:334] "Generic (PLEG): container finished" podID="43588f58-5468-4b85-8678-58bdd40823c9" containerID="6f7dea97f89e91b35d5d482bb6bba6f0beb1d46f20359e08a6a5ae4fc519776f" exitCode=0 Oct 07 21:00:01 crc kubenswrapper[4926]: I1007 21:00:01.882517 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" event={"ID":"43588f58-5468-4b85-8678-58bdd40823c9","Type":"ContainerDied","Data":"6f7dea97f89e91b35d5d482bb6bba6f0beb1d46f20359e08a6a5ae4fc519776f"} Oct 07 21:00:01 crc kubenswrapper[4926]: I1007 21:00:01.886137 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" event={"ID":"43588f58-5468-4b85-8678-58bdd40823c9","Type":"ContainerStarted","Data":"7e5a7d4e0cfed79d5450733ce3d36706640022c15f49c7a1849d62c4d7e9f0ea"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.123443 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbgt9"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.124332 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbgt9" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="registry-server" containerID="cri-o://72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4" gracePeriod=30 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.136158 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29lbg"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.136655 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-29lbg" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="registry-server" containerID="cri-o://8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2" gracePeriod=30 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.157161 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5xsvk"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.157541 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" containerID="cri-o://f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3" gracePeriod=30 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.181610 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpwnc"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.181957 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bpwnc" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="registry-server" containerID="cri-o://d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef" gracePeriod=30 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.190167 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n7ggm"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.190495 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n7ggm" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="registry-server" containerID="cri-o://ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d" gracePeriod=30 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.198272 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rt2mf"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.199299 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.202115 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rt2mf"] Oct 07 21:00:03 crc kubenswrapper[4926]: E1007 21:00:03.296278 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86367f52_f7bb_4644_ac82_c1bda5df5784.slice/crio-72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4.scope\": RecentStats: unable to find data in memory cache]" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.310927 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzpcm\" (UniqueName: \"kubernetes.io/projected/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-kube-api-access-xzpcm\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.311249 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.311281 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.311472 4926 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5xsvk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.311549 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.412777 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.412833 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.412872 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzpcm\" (UniqueName: \"kubernetes.io/projected/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-kube-api-access-xzpcm\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.414912 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.420800 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.429001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzpcm\" (UniqueName: \"kubernetes.io/projected/1ffbcf7c-a254-485e-8da4-69c2892ca0e9-kube-api-access-xzpcm\") pod \"marketplace-operator-79b997595-rt2mf\" (UID: \"1ffbcf7c-a254-485e-8da4-69c2892ca0e9\") " pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.559205 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.614872 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwn8x\" (UniqueName: \"kubernetes.io/projected/43588f58-5468-4b85-8678-58bdd40823c9-kube-api-access-wwn8x\") pod \"43588f58-5468-4b85-8678-58bdd40823c9\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.614920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/43588f58-5468-4b85-8678-58bdd40823c9-secret-volume\") pod \"43588f58-5468-4b85-8678-58bdd40823c9\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.615001 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/43588f58-5468-4b85-8678-58bdd40823c9-config-volume\") pod \"43588f58-5468-4b85-8678-58bdd40823c9\" (UID: \"43588f58-5468-4b85-8678-58bdd40823c9\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.615918 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43588f58-5468-4b85-8678-58bdd40823c9-config-volume" (OuterVolumeSpecName: "config-volume") pod "43588f58-5468-4b85-8678-58bdd40823c9" (UID: "43588f58-5468-4b85-8678-58bdd40823c9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.616006 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.618475 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.619920 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43588f58-5468-4b85-8678-58bdd40823c9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "43588f58-5468-4b85-8678-58bdd40823c9" (UID: "43588f58-5468-4b85-8678-58bdd40823c9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.621001 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43588f58-5468-4b85-8678-58bdd40823c9-kube-api-access-wwn8x" (OuterVolumeSpecName: "kube-api-access-wwn8x") pod "43588f58-5468-4b85-8678-58bdd40823c9" (UID: "43588f58-5468-4b85-8678-58bdd40823c9"). InnerVolumeSpecName "kube-api-access-wwn8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.632111 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29lbg" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.675589 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.675976 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.691861 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715645 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-catalog-content\") pod \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715702 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-catalog-content\") pod \"9035f772-a1f8-425d-90a3-00f038a08bf0\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715722 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-utilities\") pod \"9035f772-a1f8-425d-90a3-00f038a08bf0\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715762 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc9dj\" (UniqueName: \"kubernetes.io/projected/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-kube-api-access-cc9dj\") pod \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715782 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkzw5\" (UniqueName: \"kubernetes.io/projected/822a04da-8549-47c1-9994-bcc9b7ca1ed7-kube-api-access-xkzw5\") pod \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715798 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h85m8\" (UniqueName: \"kubernetes.io/projected/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-kube-api-access-h85m8\") pod \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715826 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-trusted-ca\") pod \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715844 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tssws\" (UniqueName: \"kubernetes.io/projected/86367f52-f7bb-4644-ac82-c1bda5df5784-kube-api-access-tssws\") pod \"86367f52-f7bb-4644-ac82-c1bda5df5784\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715859 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-catalog-content\") pod \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715894 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-operator-metrics\") pod \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\" (UID: \"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715922 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clc9t\" (UniqueName: \"kubernetes.io/projected/9035f772-a1f8-425d-90a3-00f038a08bf0-kube-api-access-clc9t\") pod \"9035f772-a1f8-425d-90a3-00f038a08bf0\" (UID: \"9035f772-a1f8-425d-90a3-00f038a08bf0\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715940 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-utilities\") pod \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\" (UID: \"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715966 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-utilities\") pod \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\" (UID: \"822a04da-8549-47c1-9994-bcc9b7ca1ed7\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.715983 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-utilities\") pod \"86367f52-f7bb-4644-ac82-c1bda5df5784\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.716002 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-catalog-content\") pod \"86367f52-f7bb-4644-ac82-c1bda5df5784\" (UID: \"86367f52-f7bb-4644-ac82-c1bda5df5784\") " Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.716211 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/43588f58-5468-4b85-8678-58bdd40823c9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.716223 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwn8x\" (UniqueName: \"kubernetes.io/projected/43588f58-5468-4b85-8678-58bdd40823c9-kube-api-access-wwn8x\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.716233 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/43588f58-5468-4b85-8678-58bdd40823c9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.718430 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" (UID: "d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.718377 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-utilities" (OuterVolumeSpecName: "utilities") pod "822a04da-8549-47c1-9994-bcc9b7ca1ed7" (UID: "822a04da-8549-47c1-9994-bcc9b7ca1ed7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.719376 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-utilities" (OuterVolumeSpecName: "utilities") pod "33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" (UID: "33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.719560 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-utilities" (OuterVolumeSpecName: "utilities") pod "9035f772-a1f8-425d-90a3-00f038a08bf0" (UID: "9035f772-a1f8-425d-90a3-00f038a08bf0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.719997 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9035f772-a1f8-425d-90a3-00f038a08bf0-kube-api-access-clc9t" (OuterVolumeSpecName: "kube-api-access-clc9t") pod "9035f772-a1f8-425d-90a3-00f038a08bf0" (UID: "9035f772-a1f8-425d-90a3-00f038a08bf0"). InnerVolumeSpecName "kube-api-access-clc9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.720471 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-utilities" (OuterVolumeSpecName: "utilities") pod "86367f52-f7bb-4644-ac82-c1bda5df5784" (UID: "86367f52-f7bb-4644-ac82-c1bda5df5784"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.720506 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-kube-api-access-h85m8" (OuterVolumeSpecName: "kube-api-access-h85m8") pod "33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" (UID: "33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1"). InnerVolumeSpecName "kube-api-access-h85m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.722531 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/822a04da-8549-47c1-9994-bcc9b7ca1ed7-kube-api-access-xkzw5" (OuterVolumeSpecName: "kube-api-access-xkzw5") pod "822a04da-8549-47c1-9994-bcc9b7ca1ed7" (UID: "822a04da-8549-47c1-9994-bcc9b7ca1ed7"). InnerVolumeSpecName "kube-api-access-xkzw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.723278 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-kube-api-access-cc9dj" (OuterVolumeSpecName: "kube-api-access-cc9dj") pod "d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" (UID: "d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0"). InnerVolumeSpecName "kube-api-access-cc9dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.724077 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86367f52-f7bb-4644-ac82-c1bda5df5784-kube-api-access-tssws" (OuterVolumeSpecName: "kube-api-access-tssws") pod "86367f52-f7bb-4644-ac82-c1bda5df5784" (UID: "86367f52-f7bb-4644-ac82-c1bda5df5784"). InnerVolumeSpecName "kube-api-access-tssws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.744612 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" (UID: "d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.754534 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9035f772-a1f8-425d-90a3-00f038a08bf0" (UID: "9035f772-a1f8-425d-90a3-00f038a08bf0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.790623 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86367f52-f7bb-4644-ac82-c1bda5df5784" (UID: "86367f52-f7bb-4644-ac82-c1bda5df5784"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.796166 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "822a04da-8549-47c1-9994-bcc9b7ca1ed7" (UID: "822a04da-8549-47c1-9994-bcc9b7ca1ed7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817117 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc9dj\" (UniqueName: \"kubernetes.io/projected/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-kube-api-access-cc9dj\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817148 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkzw5\" (UniqueName: \"kubernetes.io/projected/822a04da-8549-47c1-9994-bcc9b7ca1ed7-kube-api-access-xkzw5\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817158 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h85m8\" (UniqueName: \"kubernetes.io/projected/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-kube-api-access-h85m8\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817169 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817178 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tssws\" (UniqueName: \"kubernetes.io/projected/86367f52-f7bb-4644-ac82-c1bda5df5784-kube-api-access-tssws\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817187 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817210 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clc9t\" (UniqueName: \"kubernetes.io/projected/9035f772-a1f8-425d-90a3-00f038a08bf0-kube-api-access-clc9t\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817220 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817231 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817240 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817247 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86367f52-f7bb-4644-ac82-c1bda5df5784-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817257 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822a04da-8549-47c1-9994-bcc9b7ca1ed7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817268 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.817277 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9035f772-a1f8-425d-90a3-00f038a08bf0-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.826303 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" (UID: "33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.899898 4926 generic.go:334] "Generic (PLEG): container finished" podID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerID="ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d" exitCode=0 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.899971 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7ggm" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.899995 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7ggm" event={"ID":"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1","Type":"ContainerDied","Data":"ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.901354 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7ggm" event={"ID":"33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1","Type":"ContainerDied","Data":"e89e6f9d7583f93d33db7a8c57cdc8a395dafa3493525e562ecd0d167baed435"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.901403 4926 scope.go:117] "RemoveContainer" containerID="ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.904670 4926 generic.go:334] "Generic (PLEG): container finished" podID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerID="f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3" exitCode=0 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.904785 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.904965 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" event={"ID":"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0","Type":"ContainerDied","Data":"f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.905019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5xsvk" event={"ID":"d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0","Type":"ContainerDied","Data":"654c5877f62be05ff6a6661504a41a8f0486514aa3888022962bf70dac18bfa4"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.909130 4926 generic.go:334] "Generic (PLEG): container finished" podID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerID="d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef" exitCode=0 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.909252 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpwnc" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.909245 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpwnc" event={"ID":"9035f772-a1f8-425d-90a3-00f038a08bf0","Type":"ContainerDied","Data":"d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.909372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpwnc" event={"ID":"9035f772-a1f8-425d-90a3-00f038a08bf0","Type":"ContainerDied","Data":"8505377e1c8a44ea4fccc54e036ae01f82a45f0e579042301d379eee6cfbda4d"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.912877 4926 generic.go:334] "Generic (PLEG): container finished" podID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerID="72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4" exitCode=0 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.912936 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgt9" event={"ID":"86367f52-f7bb-4644-ac82-c1bda5df5784","Type":"ContainerDied","Data":"72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.912956 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgt9" event={"ID":"86367f52-f7bb-4644-ac82-c1bda5df5784","Type":"ContainerDied","Data":"d1bd551db3afcd1b1a793aa09933eb79a4ed5ca9152abf29e414fddc3d0ab777"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.913037 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgt9" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.918844 4926 generic.go:334] "Generic (PLEG): container finished" podID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerID="8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2" exitCode=0 Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.918959 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29lbg" event={"ID":"822a04da-8549-47c1-9994-bcc9b7ca1ed7","Type":"ContainerDied","Data":"8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.919009 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29lbg" event={"ID":"822a04da-8549-47c1-9994-bcc9b7ca1ed7","Type":"ContainerDied","Data":"2017d0f25c103a8d47f47609b281a22bccfb9a7b6f49b69e4ae33a78ecf11cb3"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.919857 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29lbg" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.920379 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.924952 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" event={"ID":"43588f58-5468-4b85-8678-58bdd40823c9","Type":"ContainerDied","Data":"7e5a7d4e0cfed79d5450733ce3d36706640022c15f49c7a1849d62c4d7e9f0ea"} Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.925011 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e5a7d4e0cfed79d5450733ce3d36706640022c15f49c7a1849d62c4d7e9f0ea" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.925314 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.929332 4926 scope.go:117] "RemoveContainer" containerID="0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.950660 4926 scope.go:117] "RemoveContainer" containerID="c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.961978 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n7ggm"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.972657 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n7ggm"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.985573 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5xsvk"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.992823 4926 scope.go:117] "RemoveContainer" containerID="ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.992958 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5xsvk"] Oct 07 21:00:03 crc kubenswrapper[4926]: E1007 21:00:03.993530 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d\": container with ID starting with ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d not found: ID does not exist" containerID="ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.993677 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d"} err="failed to get container status \"ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d\": rpc error: code = NotFound desc = could not find container \"ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d\": container with ID starting with ba0ae7d9a78a9e08fd7bec897a323d9ae0783e0eb64302987f9d82c6b81ad10d not found: ID does not exist" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.993710 4926 scope.go:117] "RemoveContainer" containerID="0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5" Oct 07 21:00:03 crc kubenswrapper[4926]: E1007 21:00:03.993926 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5\": container with ID starting with 0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5 not found: ID does not exist" containerID="0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.993948 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5"} err="failed to get container status \"0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5\": rpc error: code = NotFound desc = could not find container \"0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5\": container with ID starting with 0eef9119f343236f052205c83453833b8265c67b82d9747c3cd3ac054cec89f5 not found: ID does not exist" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.993962 4926 scope.go:117] "RemoveContainer" containerID="c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741" Oct 07 21:00:03 crc kubenswrapper[4926]: E1007 21:00:03.994212 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741\": container with ID starting with c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741 not found: ID does not exist" containerID="c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.994252 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741"} err="failed to get container status \"c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741\": rpc error: code = NotFound desc = could not find container \"c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741\": container with ID starting with c1d4360bf44d4834f7687459b24e1ec5dedce4cea5758218fd6cbc65ece8c741 not found: ID does not exist" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.994267 4926 scope.go:117] "RemoveContainer" containerID="f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3" Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.995836 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29lbg"] Oct 07 21:00:03 crc kubenswrapper[4926]: I1007 21:00:03.998451 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-29lbg"] Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.001931 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpwnc"] Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.007602 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpwnc"] Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.007669 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbgt9"] Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.008603 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbgt9"] Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.011502 4926 scope.go:117] "RemoveContainer" containerID="f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.012035 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3\": container with ID starting with f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3 not found: ID does not exist" containerID="f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.012071 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3"} err="failed to get container status \"f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3\": rpc error: code = NotFound desc = could not find container \"f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3\": container with ID starting with f406e16bda6c52a0abd503867cee7501699def6b3a6a250569e759c7634c09f3 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.012105 4926 scope.go:117] "RemoveContainer" containerID="d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.030344 4926 scope.go:117] "RemoveContainer" containerID="821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.041610 4926 scope.go:117] "RemoveContainer" containerID="10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.051476 4926 scope.go:117] "RemoveContainer" containerID="d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.051771 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef\": container with ID starting with d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef not found: ID does not exist" containerID="d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.051832 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef"} err="failed to get container status \"d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef\": rpc error: code = NotFound desc = could not find container \"d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef\": container with ID starting with d7a2436bcf33ab17eea490c204ccde646a2c9b0762227497347923278224a9ef not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.051868 4926 scope.go:117] "RemoveContainer" containerID="821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.052146 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1\": container with ID starting with 821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1 not found: ID does not exist" containerID="821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.052211 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1"} err="failed to get container status \"821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1\": rpc error: code = NotFound desc = could not find container \"821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1\": container with ID starting with 821f94d8f54035866acad987ba1f291aaafb9e0c41d05a6d99920ce9c51b21a1 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.052247 4926 scope.go:117] "RemoveContainer" containerID="10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.052530 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70\": container with ID starting with 10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70 not found: ID does not exist" containerID="10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.052567 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70"} err="failed to get container status \"10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70\": rpc error: code = NotFound desc = could not find container \"10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70\": container with ID starting with 10b1e05c3a9f1186b355b4f6009c326e458da4b8308c8dd013f696fe4dfa9d70 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.052589 4926 scope.go:117] "RemoveContainer" containerID="72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.061505 4926 scope.go:117] "RemoveContainer" containerID="2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.075308 4926 scope.go:117] "RemoveContainer" containerID="4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.085994 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rt2mf"] Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.087932 4926 scope.go:117] "RemoveContainer" containerID="72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.088251 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4\": container with ID starting with 72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4 not found: ID does not exist" containerID="72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.088285 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4"} err="failed to get container status \"72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4\": rpc error: code = NotFound desc = could not find container \"72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4\": container with ID starting with 72fc2297f0a1f30d72ea3a6a1c1cb560a684b1d6a40d4faeeac48f01532ce2a4 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.088310 4926 scope.go:117] "RemoveContainer" containerID="2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.088561 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a\": container with ID starting with 2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a not found: ID does not exist" containerID="2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.088608 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a"} err="failed to get container status \"2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a\": rpc error: code = NotFound desc = could not find container \"2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a\": container with ID starting with 2ab94790ac5d3163e7fd4528c70872ee0f81083a15f859a3178c0103a104d02a not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.088641 4926 scope.go:117] "RemoveContainer" containerID="4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.088900 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3\": container with ID starting with 4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3 not found: ID does not exist" containerID="4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.088928 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3"} err="failed to get container status \"4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3\": rpc error: code = NotFound desc = could not find container \"4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3\": container with ID starting with 4f945c0851d24c3dd0eb25c4151bfb05bdf868d5152cfa4197130cccd0e24ae3 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.088945 4926 scope.go:117] "RemoveContainer" containerID="8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.100414 4926 scope.go:117] "RemoveContainer" containerID="a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.120703 4926 scope.go:117] "RemoveContainer" containerID="5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.143670 4926 scope.go:117] "RemoveContainer" containerID="8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.144379 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2\": container with ID starting with 8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2 not found: ID does not exist" containerID="8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.144459 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2"} err="failed to get container status \"8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2\": rpc error: code = NotFound desc = could not find container \"8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2\": container with ID starting with 8f4c2bf0c6f84f23e18b5bd8b5072e31fb959ae034a806d271ea0f5f055013e2 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.144520 4926 scope.go:117] "RemoveContainer" containerID="a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.145421 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd\": container with ID starting with a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd not found: ID does not exist" containerID="a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.145476 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd"} err="failed to get container status \"a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd\": rpc error: code = NotFound desc = could not find container \"a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd\": container with ID starting with a330ac9e1b2489dfdecca7344d7ab7f70fae90d343a8e561304a6695a07123dd not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.145512 4926 scope.go:117] "RemoveContainer" containerID="5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12" Oct 07 21:00:04 crc kubenswrapper[4926]: E1007 21:00:04.145809 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12\": container with ID starting with 5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12 not found: ID does not exist" containerID="5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.145836 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12"} err="failed to get container status \"5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12\": rpc error: code = NotFound desc = could not find container \"5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12\": container with ID starting with 5532b5bbf1ffb105145126e37435894ece45f62756a8f15a936b9a0ced273a12 not found: ID does not exist" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.685939 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" path="/var/lib/kubelet/pods/33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1/volumes" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.687569 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" path="/var/lib/kubelet/pods/822a04da-8549-47c1-9994-bcc9b7ca1ed7/volumes" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.688814 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" path="/var/lib/kubelet/pods/86367f52-f7bb-4644-ac82-c1bda5df5784/volumes" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.691494 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" path="/var/lib/kubelet/pods/9035f772-a1f8-425d-90a3-00f038a08bf0/volumes" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.692926 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" path="/var/lib/kubelet/pods/d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0/volumes" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.934352 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" event={"ID":"1ffbcf7c-a254-485e-8da4-69c2892ca0e9","Type":"ContainerStarted","Data":"9a4f9fbd133ed8586389dcfb0f9e3df7f835c4798a136067e7fb85c38f797b97"} Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.934390 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" event={"ID":"1ffbcf7c-a254-485e-8da4-69c2892ca0e9","Type":"ContainerStarted","Data":"ba82ac50ef27342f224fe8112746611058d1c5bceb3909f092c8f8cbca649a68"} Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.935374 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.938087 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" Oct 07 21:00:04 crc kubenswrapper[4926]: I1007 21:00:04.970528 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rt2mf" podStartSLOduration=1.970511793 podStartE2EDuration="1.970511793s" podCreationTimestamp="2025-10-07 21:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:00:04.949570685 +0000 UTC m=+254.988151835" watchObservedRunningTime="2025-10-07 21:00:04.970511793 +0000 UTC m=+255.009092943" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339492 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ncl5q"] Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339735 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339750 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339762 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339770 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339781 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339791 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339801 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339808 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339824 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339831 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339840 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43588f58-5468-4b85-8678-58bdd40823c9" containerName="collect-profiles" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339847 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="43588f58-5468-4b85-8678-58bdd40823c9" containerName="collect-profiles" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339856 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339863 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339875 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339883 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339894 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339902 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339913 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339923 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339934 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339942 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339953 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339962 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="extract-utilities" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339969 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339977 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: E1007 21:00:05.339986 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.339995 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="extract-content" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.340106 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="43588f58-5468-4b85-8678-58bdd40823c9" containerName="collect-profiles" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.340124 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="86367f52-f7bb-4644-ac82-c1bda5df5784" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.340135 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9035f772-a1f8-425d-90a3-00f038a08bf0" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.340147 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ec3fd9-fd3c-4d2a-9293-b7eac9b14ab1" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.340214 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="822a04da-8549-47c1-9994-bcc9b7ca1ed7" containerName="registry-server" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.340222 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a51600-14eb-4bd3-8ff3-cf9fa4764bc0" containerName="marketplace-operator" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.341035 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.344581 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.351111 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ncl5q"] Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.433730 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-utilities\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.433766 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5zrq\" (UniqueName: \"kubernetes.io/projected/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-kube-api-access-b5zrq\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.433806 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-catalog-content\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.535481 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-catalog-content\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.535591 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-utilities\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.535614 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5zrq\" (UniqueName: \"kubernetes.io/projected/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-kube-api-access-b5zrq\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.536454 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-catalog-content\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.536578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-utilities\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.539853 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dq9zz"] Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.540824 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.551968 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.555837 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dq9zz"] Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.562860 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5zrq\" (UniqueName: \"kubernetes.io/projected/319f1329-c2a8-40eb-a8aa-bf7c4347fe47-kube-api-access-b5zrq\") pod \"redhat-marketplace-ncl5q\" (UID: \"319f1329-c2a8-40eb-a8aa-bf7c4347fe47\") " pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.637266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zq5c\" (UniqueName: \"kubernetes.io/projected/eb0efddf-5b03-4146-8ecb-ee37576f129c-kube-api-access-8zq5c\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.637369 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0efddf-5b03-4146-8ecb-ee37576f129c-utilities\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.637401 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0efddf-5b03-4146-8ecb-ee37576f129c-catalog-content\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.661947 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.739820 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zq5c\" (UniqueName: \"kubernetes.io/projected/eb0efddf-5b03-4146-8ecb-ee37576f129c-kube-api-access-8zq5c\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.739899 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0efddf-5b03-4146-8ecb-ee37576f129c-utilities\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.739933 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0efddf-5b03-4146-8ecb-ee37576f129c-catalog-content\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.740405 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0efddf-5b03-4146-8ecb-ee37576f129c-utilities\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.740489 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0efddf-5b03-4146-8ecb-ee37576f129c-catalog-content\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.765315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zq5c\" (UniqueName: \"kubernetes.io/projected/eb0efddf-5b03-4146-8ecb-ee37576f129c-kube-api-access-8zq5c\") pod \"redhat-operators-dq9zz\" (UID: \"eb0efddf-5b03-4146-8ecb-ee37576f129c\") " pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:05 crc kubenswrapper[4926]: I1007 21:00:05.864655 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.083187 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ncl5q"] Oct 07 21:00:06 crc kubenswrapper[4926]: W1007 21:00:06.089548 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod319f1329_c2a8_40eb_a8aa_bf7c4347fe47.slice/crio-8c68f78d32344725586f86446bf28bd69cd8b52b09d43bc5c0c178f7131c5892 WatchSource:0}: Error finding container 8c68f78d32344725586f86446bf28bd69cd8b52b09d43bc5c0c178f7131c5892: Status 404 returned error can't find the container with id 8c68f78d32344725586f86446bf28bd69cd8b52b09d43bc5c0c178f7131c5892 Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.282438 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dq9zz"] Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.954817 4926 generic.go:334] "Generic (PLEG): container finished" podID="319f1329-c2a8-40eb-a8aa-bf7c4347fe47" containerID="7562b25821f173f4c091514febfb12baa9aa127209d621c93769eef33d55d325" exitCode=0 Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.954940 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncl5q" event={"ID":"319f1329-c2a8-40eb-a8aa-bf7c4347fe47","Type":"ContainerDied","Data":"7562b25821f173f4c091514febfb12baa9aa127209d621c93769eef33d55d325"} Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.955985 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncl5q" event={"ID":"319f1329-c2a8-40eb-a8aa-bf7c4347fe47","Type":"ContainerStarted","Data":"8c68f78d32344725586f86446bf28bd69cd8b52b09d43bc5c0c178f7131c5892"} Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.962390 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb0efddf-5b03-4146-8ecb-ee37576f129c" containerID="785fbc1691ae0abb286080c3436bc2e841c3bd69b7a90a013faf9fa847885533" exitCode=0 Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.963426 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq9zz" event={"ID":"eb0efddf-5b03-4146-8ecb-ee37576f129c","Type":"ContainerDied","Data":"785fbc1691ae0abb286080c3436bc2e841c3bd69b7a90a013faf9fa847885533"} Oct 07 21:00:06 crc kubenswrapper[4926]: I1007 21:00:06.963458 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq9zz" event={"ID":"eb0efddf-5b03-4146-8ecb-ee37576f129c","Type":"ContainerStarted","Data":"539b3347b61c01bdc9c2bb0816885b2cc05900d24d6e3c9a96a1fbd4b0197dcd"} Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.744136 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rbmzv"] Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.748961 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.753353 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.756419 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rbmzv"] Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.767405 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmdzg\" (UniqueName: \"kubernetes.io/projected/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-kube-api-access-gmdzg\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.767728 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-utilities\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.767823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-catalog-content\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.869443 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmdzg\" (UniqueName: \"kubernetes.io/projected/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-kube-api-access-gmdzg\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.869773 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-utilities\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.869815 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-catalog-content\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.870329 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-catalog-content\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.872576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-utilities\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.900124 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmdzg\" (UniqueName: \"kubernetes.io/projected/261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509-kube-api-access-gmdzg\") pod \"certified-operators-rbmzv\" (UID: \"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509\") " pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.940053 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5p9gr"] Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.942287 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.945244 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.950446 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5p9gr"] Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.970951 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgq5t\" (UniqueName: \"kubernetes.io/projected/e3e61957-dad4-4160-8916-672ce8467461-kube-api-access-pgq5t\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.970984 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3e61957-dad4-4160-8916-672ce8467461-catalog-content\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:07 crc kubenswrapper[4926]: I1007 21:00:07.971008 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3e61957-dad4-4160-8916-672ce8467461-utilities\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.072032 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgq5t\" (UniqueName: \"kubernetes.io/projected/e3e61957-dad4-4160-8916-672ce8467461-kube-api-access-pgq5t\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.072304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3e61957-dad4-4160-8916-672ce8467461-catalog-content\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.072445 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3e61957-dad4-4160-8916-672ce8467461-utilities\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.072761 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3e61957-dad4-4160-8916-672ce8467461-catalog-content\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.072891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3e61957-dad4-4160-8916-672ce8467461-utilities\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.086015 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.094699 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgq5t\" (UniqueName: \"kubernetes.io/projected/e3e61957-dad4-4160-8916-672ce8467461-kube-api-access-pgq5t\") pod \"community-operators-5p9gr\" (UID: \"e3e61957-dad4-4160-8916-672ce8467461\") " pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.301410 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.474370 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rbmzv"] Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.724418 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5p9gr"] Oct 07 21:00:08 crc kubenswrapper[4926]: W1007 21:00:08.726524 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3e61957_dad4_4160_8916_672ce8467461.slice/crio-ac5e735cffd55e2235307c1a73135d6f62c4540b767b137356bb6f096bff4db9 WatchSource:0}: Error finding container ac5e735cffd55e2235307c1a73135d6f62c4540b767b137356bb6f096bff4db9: Status 404 returned error can't find the container with id ac5e735cffd55e2235307c1a73135d6f62c4540b767b137356bb6f096bff4db9 Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.982659 4926 generic.go:334] "Generic (PLEG): container finished" podID="261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509" containerID="697b9355803c4f7e90a8c1dddb961ce638a46c3da4d5a064d5b80908bcd98b1c" exitCode=0 Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.982848 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbmzv" event={"ID":"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509","Type":"ContainerDied","Data":"697b9355803c4f7e90a8c1dddb961ce638a46c3da4d5a064d5b80908bcd98b1c"} Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.983077 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbmzv" event={"ID":"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509","Type":"ContainerStarted","Data":"072f67a057d93f38905ea443d9fbc6b805acd07d9d0ced296a65246a49c66066"} Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.985307 4926 generic.go:334] "Generic (PLEG): container finished" podID="e3e61957-dad4-4160-8916-672ce8467461" containerID="4a09f7ab32e5378c02e2ca8f04a3415404765d6fd89d0f4092bf767336cb1e82" exitCode=0 Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.985415 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p9gr" event={"ID":"e3e61957-dad4-4160-8916-672ce8467461","Type":"ContainerDied","Data":"4a09f7ab32e5378c02e2ca8f04a3415404765d6fd89d0f4092bf767336cb1e82"} Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.985487 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p9gr" event={"ID":"e3e61957-dad4-4160-8916-672ce8467461","Type":"ContainerStarted","Data":"ac5e735cffd55e2235307c1a73135d6f62c4540b767b137356bb6f096bff4db9"} Oct 07 21:00:08 crc kubenswrapper[4926]: I1007 21:00:08.999682 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb0efddf-5b03-4146-8ecb-ee37576f129c" containerID="e691789eff030310a67c8f8c0e410d4782424eb867f30c08e935cac7c0a7f816" exitCode=0 Oct 07 21:00:09 crc kubenswrapper[4926]: I1007 21:00:09.000221 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq9zz" event={"ID":"eb0efddf-5b03-4146-8ecb-ee37576f129c","Type":"ContainerDied","Data":"e691789eff030310a67c8f8c0e410d4782424eb867f30c08e935cac7c0a7f816"} Oct 07 21:00:09 crc kubenswrapper[4926]: I1007 21:00:09.003468 4926 generic.go:334] "Generic (PLEG): container finished" podID="319f1329-c2a8-40eb-a8aa-bf7c4347fe47" containerID="4744baaef66b26c882ed2eb5945bd51c5e085b9c4988f95b9a25077bff2adedf" exitCode=0 Oct 07 21:00:09 crc kubenswrapper[4926]: I1007 21:00:09.003531 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncl5q" event={"ID":"319f1329-c2a8-40eb-a8aa-bf7c4347fe47","Type":"ContainerDied","Data":"4744baaef66b26c882ed2eb5945bd51c5e085b9c4988f95b9a25077bff2adedf"} Oct 07 21:00:10 crc kubenswrapper[4926]: I1007 21:00:10.018895 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq9zz" event={"ID":"eb0efddf-5b03-4146-8ecb-ee37576f129c","Type":"ContainerStarted","Data":"9c6049dd3bdf2e44d36bbc28030ae9ffba61acc01a26146d053ce4076aa38524"} Oct 07 21:00:10 crc kubenswrapper[4926]: I1007 21:00:10.026469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncl5q" event={"ID":"319f1329-c2a8-40eb-a8aa-bf7c4347fe47","Type":"ContainerStarted","Data":"be77f61d852275af41fca773d8f0d3e258441e12886fb4325f60b4770f4785ee"} Oct 07 21:00:10 crc kubenswrapper[4926]: I1007 21:00:10.028855 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbmzv" event={"ID":"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509","Type":"ContainerStarted","Data":"74998d9b70b4817f2dc1e7dbf1d0e84787c5a1ca86fa54dfbaf1be090458f95f"} Oct 07 21:00:10 crc kubenswrapper[4926]: I1007 21:00:10.035057 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dq9zz" podStartSLOduration=2.409241291 podStartE2EDuration="5.035029757s" podCreationTimestamp="2025-10-07 21:00:05 +0000 UTC" firstStartedPulling="2025-10-07 21:00:06.963861 +0000 UTC m=+257.002442150" lastFinishedPulling="2025-10-07 21:00:09.589649456 +0000 UTC m=+259.628230616" observedRunningTime="2025-10-07 21:00:10.033290935 +0000 UTC m=+260.071872105" watchObservedRunningTime="2025-10-07 21:00:10.035029757 +0000 UTC m=+260.073610917" Oct 07 21:00:10 crc kubenswrapper[4926]: I1007 21:00:10.055258 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ncl5q" podStartSLOduration=2.55795485 podStartE2EDuration="5.055236723s" podCreationTimestamp="2025-10-07 21:00:05 +0000 UTC" firstStartedPulling="2025-10-07 21:00:06.958511127 +0000 UTC m=+256.997092277" lastFinishedPulling="2025-10-07 21:00:09.455793 +0000 UTC m=+259.494374150" observedRunningTime="2025-10-07 21:00:10.051710195 +0000 UTC m=+260.090291355" watchObservedRunningTime="2025-10-07 21:00:10.055236723 +0000 UTC m=+260.093817873" Oct 07 21:00:11 crc kubenswrapper[4926]: I1007 21:00:11.038957 4926 generic.go:334] "Generic (PLEG): container finished" podID="261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509" containerID="74998d9b70b4817f2dc1e7dbf1d0e84787c5a1ca86fa54dfbaf1be090458f95f" exitCode=0 Oct 07 21:00:11 crc kubenswrapper[4926]: I1007 21:00:11.039021 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbmzv" event={"ID":"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509","Type":"ContainerDied","Data":"74998d9b70b4817f2dc1e7dbf1d0e84787c5a1ca86fa54dfbaf1be090458f95f"} Oct 07 21:00:12 crc kubenswrapper[4926]: I1007 21:00:12.045737 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbmzv" event={"ID":"261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509","Type":"ContainerStarted","Data":"faaf7b4e3bac437fcc98d538675619122ca04d9bcff8fba909585fc7bbcf60f2"} Oct 07 21:00:12 crc kubenswrapper[4926]: I1007 21:00:12.047674 4926 generic.go:334] "Generic (PLEG): container finished" podID="e3e61957-dad4-4160-8916-672ce8467461" containerID="e76656e20e1f085e0a7d7e7fd1cfe73886e8b6af403f43eadafe9fc3be6ce344" exitCode=0 Oct 07 21:00:12 crc kubenswrapper[4926]: I1007 21:00:12.047700 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p9gr" event={"ID":"e3e61957-dad4-4160-8916-672ce8467461","Type":"ContainerDied","Data":"e76656e20e1f085e0a7d7e7fd1cfe73886e8b6af403f43eadafe9fc3be6ce344"} Oct 07 21:00:12 crc kubenswrapper[4926]: I1007 21:00:12.065397 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rbmzv" podStartSLOduration=2.3812799350000002 podStartE2EDuration="5.065384073s" podCreationTimestamp="2025-10-07 21:00:07 +0000 UTC" firstStartedPulling="2025-10-07 21:00:08.984608163 +0000 UTC m=+259.023189333" lastFinishedPulling="2025-10-07 21:00:11.668712281 +0000 UTC m=+261.707293471" observedRunningTime="2025-10-07 21:00:12.064290011 +0000 UTC m=+262.102871161" watchObservedRunningTime="2025-10-07 21:00:12.065384073 +0000 UTC m=+262.103965223" Oct 07 21:00:13 crc kubenswrapper[4926]: I1007 21:00:13.054784 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p9gr" event={"ID":"e3e61957-dad4-4160-8916-672ce8467461","Type":"ContainerStarted","Data":"01604ea7ff153621fcee7a94584cc00c52e4921606c854fe0899d07c4b6c2528"} Oct 07 21:00:13 crc kubenswrapper[4926]: I1007 21:00:13.075904 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5p9gr" podStartSLOduration=2.634449639 podStartE2EDuration="6.075885948s" podCreationTimestamp="2025-10-07 21:00:07 +0000 UTC" firstStartedPulling="2025-10-07 21:00:08.997788114 +0000 UTC m=+259.036369264" lastFinishedPulling="2025-10-07 21:00:12.439224423 +0000 UTC m=+262.477805573" observedRunningTime="2025-10-07 21:00:13.072648011 +0000 UTC m=+263.111229171" watchObservedRunningTime="2025-10-07 21:00:13.075885948 +0000 UTC m=+263.114467108" Oct 07 21:00:15 crc kubenswrapper[4926]: I1007 21:00:15.662207 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:15 crc kubenswrapper[4926]: I1007 21:00:15.662580 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:15 crc kubenswrapper[4926]: I1007 21:00:15.730415 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:15 crc kubenswrapper[4926]: I1007 21:00:15.865387 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:15 crc kubenswrapper[4926]: I1007 21:00:15.865437 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:15 crc kubenswrapper[4926]: I1007 21:00:15.908328 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:16 crc kubenswrapper[4926]: I1007 21:00:16.119414 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dq9zz" Oct 07 21:00:16 crc kubenswrapper[4926]: I1007 21:00:16.131239 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ncl5q" Oct 07 21:00:18 crc kubenswrapper[4926]: I1007 21:00:18.086964 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:18 crc kubenswrapper[4926]: I1007 21:00:18.087390 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:18 crc kubenswrapper[4926]: I1007 21:00:18.128214 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:00:18 crc kubenswrapper[4926]: I1007 21:00:18.303043 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:18 crc kubenswrapper[4926]: I1007 21:00:18.303088 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:18 crc kubenswrapper[4926]: I1007 21:00:18.337597 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:19 crc kubenswrapper[4926]: I1007 21:00:19.127879 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5p9gr" Oct 07 21:00:19 crc kubenswrapper[4926]: I1007 21:00:19.140685 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rbmzv" Oct 07 21:01:33 crc kubenswrapper[4926]: I1007 21:01:33.211826 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:01:33 crc kubenswrapper[4926]: I1007 21:01:33.212272 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:02:03 crc kubenswrapper[4926]: I1007 21:02:03.211164 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:02:03 crc kubenswrapper[4926]: I1007 21:02:03.211905 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.465343 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-v647p"] Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.466846 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.506011 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-v647p"] Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.513230 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-registry-tls\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.514105 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.514241 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f8f22ba9-47de-4e3a-995e-674e268222bb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.514336 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f8f22ba9-47de-4e3a-995e-674e268222bb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.514457 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-bound-sa-token\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.514562 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k7wp\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-kube-api-access-2k7wp\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.515600 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8f22ba9-47de-4e3a-995e-674e268222bb-trusted-ca\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.515774 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f8f22ba9-47de-4e3a-995e-674e268222bb-registry-certificates\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.548321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.616846 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f8f22ba9-47de-4e3a-995e-674e268222bb-registry-certificates\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.616903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-registry-tls\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.616925 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f8f22ba9-47de-4e3a-995e-674e268222bb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.616948 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f8f22ba9-47de-4e3a-995e-674e268222bb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.616984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-bound-sa-token\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.617010 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k7wp\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-kube-api-access-2k7wp\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.617026 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8f22ba9-47de-4e3a-995e-674e268222bb-trusted-ca\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.617857 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f8f22ba9-47de-4e3a-995e-674e268222bb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.618372 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8f22ba9-47de-4e3a-995e-674e268222bb-trusted-ca\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.619584 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f8f22ba9-47de-4e3a-995e-674e268222bb-registry-certificates\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.624009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f8f22ba9-47de-4e3a-995e-674e268222bb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.624306 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-registry-tls\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.635708 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k7wp\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-kube-api-access-2k7wp\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.639283 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8f22ba9-47de-4e3a-995e-674e268222bb-bound-sa-token\") pod \"image-registry-66df7c8f76-v647p\" (UID: \"f8f22ba9-47de-4e3a-995e-674e268222bb\") " pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:19 crc kubenswrapper[4926]: I1007 21:02:19.793316 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:20 crc kubenswrapper[4926]: I1007 21:02:20.112495 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-v647p"] Oct 07 21:02:20 crc kubenswrapper[4926]: W1007 21:02:20.126851 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8f22ba9_47de_4e3a_995e_674e268222bb.slice/crio-ef1e4b04279bb575826bc065c752e7349b59a311fdf8925a84f9ba94c49b6fd6 WatchSource:0}: Error finding container ef1e4b04279bb575826bc065c752e7349b59a311fdf8925a84f9ba94c49b6fd6: Status 404 returned error can't find the container with id ef1e4b04279bb575826bc065c752e7349b59a311fdf8925a84f9ba94c49b6fd6 Oct 07 21:02:20 crc kubenswrapper[4926]: I1007 21:02:20.893451 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" event={"ID":"f8f22ba9-47de-4e3a-995e-674e268222bb","Type":"ContainerStarted","Data":"624f130679bb4c91e163e088c3111a5a385b01bdbd351c5843e2019b4f1d5e76"} Oct 07 21:02:20 crc kubenswrapper[4926]: I1007 21:02:20.893829 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:20 crc kubenswrapper[4926]: I1007 21:02:20.893859 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" event={"ID":"f8f22ba9-47de-4e3a-995e-674e268222bb","Type":"ContainerStarted","Data":"ef1e4b04279bb575826bc065c752e7349b59a311fdf8925a84f9ba94c49b6fd6"} Oct 07 21:02:20 crc kubenswrapper[4926]: I1007 21:02:20.919363 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" podStartSLOduration=1.919347704 podStartE2EDuration="1.919347704s" podCreationTimestamp="2025-10-07 21:02:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:02:20.916275433 +0000 UTC m=+390.954856593" watchObservedRunningTime="2025-10-07 21:02:20.919347704 +0000 UTC m=+390.957928854" Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.209989 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.210796 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.210921 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.211714 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03c2adc15ef42fc76b9b086cb5172f247b75e21cd7f7824690f65b2c4a8e8842"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.211774 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://03c2adc15ef42fc76b9b086cb5172f247b75e21cd7f7824690f65b2c4a8e8842" gracePeriod=600 Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.983406 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="03c2adc15ef42fc76b9b086cb5172f247b75e21cd7f7824690f65b2c4a8e8842" exitCode=0 Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.983477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"03c2adc15ef42fc76b9b086cb5172f247b75e21cd7f7824690f65b2c4a8e8842"} Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.984278 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"ea4910515b9deac1bdd55447e1b44ec9155b5c86397c1d6d38253d2a05662255"} Oct 07 21:02:33 crc kubenswrapper[4926]: I1007 21:02:33.984310 4926 scope.go:117] "RemoveContainer" containerID="bf0b6d063e1bdbddc895b490260d741f0383a29e0fcbe71f56769a8f39496327" Oct 07 21:02:39 crc kubenswrapper[4926]: I1007 21:02:39.799626 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-v647p" Oct 07 21:02:39 crc kubenswrapper[4926]: I1007 21:02:39.871079 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kfjx5"] Oct 07 21:03:04 crc kubenswrapper[4926]: I1007 21:03:04.928542 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" podUID="1ec7685a-e96e-434f-be3a-cda923c04ba1" containerName="registry" containerID="cri-o://ea2bf0b3bc9707d468947c4f6ab9b0fbfa3a88740a287d98991f49dd0364ffeb" gracePeriod=30 Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.206801 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ec7685a-e96e-434f-be3a-cda923c04ba1" containerID="ea2bf0b3bc9707d468947c4f6ab9b0fbfa3a88740a287d98991f49dd0364ffeb" exitCode=0 Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.207296 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" event={"ID":"1ec7685a-e96e-434f-be3a-cda923c04ba1","Type":"ContainerDied","Data":"ea2bf0b3bc9707d468947c4f6ab9b0fbfa3a88740a287d98991f49dd0364ffeb"} Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.407126 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.519926 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-bound-sa-token\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.519979 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-certificates\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.520030 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-trusted-ca\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.520175 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.520240 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-tls\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.520267 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ec7685a-e96e-434f-be3a-cda923c04ba1-installation-pull-secrets\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.520346 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9rhp\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-kube-api-access-p9rhp\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.520374 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ec7685a-e96e-434f-be3a-cda923c04ba1-ca-trust-extracted\") pod \"1ec7685a-e96e-434f-be3a-cda923c04ba1\" (UID: \"1ec7685a-e96e-434f-be3a-cda923c04ba1\") " Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.521429 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.521624 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.526676 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.527484 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ec7685a-e96e-434f-be3a-cda923c04ba1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.528104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.529030 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-kube-api-access-p9rhp" (OuterVolumeSpecName: "kube-api-access-p9rhp") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "kube-api-access-p9rhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.536870 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.554364 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec7685a-e96e-434f-be3a-cda923c04ba1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1ec7685a-e96e-434f-be3a-cda923c04ba1" (UID: "1ec7685a-e96e-434f-be3a-cda923c04ba1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621828 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621878 4926 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621898 4926 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ec7685a-e96e-434f-be3a-cda923c04ba1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621921 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9rhp\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-kube-api-access-p9rhp\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621938 4926 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ec7685a-e96e-434f-be3a-cda923c04ba1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621955 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ec7685a-e96e-434f-be3a-cda923c04ba1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:05 crc kubenswrapper[4926]: I1007 21:03:05.621972 4926 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ec7685a-e96e-434f-be3a-cda923c04ba1-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 21:03:06 crc kubenswrapper[4926]: I1007 21:03:06.214712 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" event={"ID":"1ec7685a-e96e-434f-be3a-cda923c04ba1","Type":"ContainerDied","Data":"c786f57d42b2379303ec14bb07af640a5a553c6b4dd9adb7d9b3596992ebedea"} Oct 07 21:03:06 crc kubenswrapper[4926]: I1007 21:03:06.215026 4926 scope.go:117] "RemoveContainer" containerID="ea2bf0b3bc9707d468947c4f6ab9b0fbfa3a88740a287d98991f49dd0364ffeb" Oct 07 21:03:06 crc kubenswrapper[4926]: I1007 21:03:06.214796 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kfjx5" Oct 07 21:03:06 crc kubenswrapper[4926]: I1007 21:03:06.261505 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kfjx5"] Oct 07 21:03:06 crc kubenswrapper[4926]: I1007 21:03:06.267360 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kfjx5"] Oct 07 21:03:06 crc kubenswrapper[4926]: I1007 21:03:06.690277 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec7685a-e96e-434f-be3a-cda923c04ba1" path="/var/lib/kubelet/pods/1ec7685a-e96e-434f-be3a-cda923c04ba1/volumes" Oct 07 21:04:33 crc kubenswrapper[4926]: I1007 21:04:33.210866 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:04:33 crc kubenswrapper[4926]: I1007 21:04:33.211508 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:05:03 crc kubenswrapper[4926]: I1007 21:05:03.209786 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:05:03 crc kubenswrapper[4926]: I1007 21:05:03.210578 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.206076 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-77j9j"] Oct 07 21:05:10 crc kubenswrapper[4926]: E1007 21:05:10.207273 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec7685a-e96e-434f-be3a-cda923c04ba1" containerName="registry" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.207289 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec7685a-e96e-434f-be3a-cda923c04ba1" containerName="registry" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.207413 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec7685a-e96e-434f-be3a-cda923c04ba1" containerName="registry" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.207901 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.209700 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.210161 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-67xk2" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.210172 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.218435 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-77j9j"] Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.223380 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-v56pb"] Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.224385 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-v56pb" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.231117 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-ggf75" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.243167 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dd6ts"] Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.243991 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.246690 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-x75wp" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.248764 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-v56pb"] Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.282239 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dd6ts"] Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.332256 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s86kf\" (UniqueName: \"kubernetes.io/projected/742ac735-fc39-4ab6-a072-834db359b8d8-kube-api-access-s86kf\") pod \"cert-manager-5b446d88c5-v56pb\" (UID: \"742ac735-fc39-4ab6-a072-834db359b8d8\") " pod="cert-manager/cert-manager-5b446d88c5-v56pb" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.332321 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4dg8\" (UniqueName: \"kubernetes.io/projected/3aa5ce8c-a725-4bd2-a504-887c768a64b6-kube-api-access-c4dg8\") pod \"cert-manager-cainjector-7f985d654d-77j9j\" (UID: \"3aa5ce8c-a725-4bd2-a504-887c768a64b6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.433867 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6wwq\" (UniqueName: \"kubernetes.io/projected/07446e4f-68a2-422e-b303-e3637d956a4c-kube-api-access-p6wwq\") pod \"cert-manager-webhook-5655c58dd6-dd6ts\" (UID: \"07446e4f-68a2-422e-b303-e3637d956a4c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.433990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s86kf\" (UniqueName: \"kubernetes.io/projected/742ac735-fc39-4ab6-a072-834db359b8d8-kube-api-access-s86kf\") pod \"cert-manager-5b446d88c5-v56pb\" (UID: \"742ac735-fc39-4ab6-a072-834db359b8d8\") " pod="cert-manager/cert-manager-5b446d88c5-v56pb" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.434019 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4dg8\" (UniqueName: \"kubernetes.io/projected/3aa5ce8c-a725-4bd2-a504-887c768a64b6-kube-api-access-c4dg8\") pod \"cert-manager-cainjector-7f985d654d-77j9j\" (UID: \"3aa5ce8c-a725-4bd2-a504-887c768a64b6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.465213 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s86kf\" (UniqueName: \"kubernetes.io/projected/742ac735-fc39-4ab6-a072-834db359b8d8-kube-api-access-s86kf\") pod \"cert-manager-5b446d88c5-v56pb\" (UID: \"742ac735-fc39-4ab6-a072-834db359b8d8\") " pod="cert-manager/cert-manager-5b446d88c5-v56pb" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.465224 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4dg8\" (UniqueName: \"kubernetes.io/projected/3aa5ce8c-a725-4bd2-a504-887c768a64b6-kube-api-access-c4dg8\") pod \"cert-manager-cainjector-7f985d654d-77j9j\" (UID: \"3aa5ce8c-a725-4bd2-a504-887c768a64b6\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.534936 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.535617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6wwq\" (UniqueName: \"kubernetes.io/projected/07446e4f-68a2-422e-b303-e3637d956a4c-kube-api-access-p6wwq\") pod \"cert-manager-webhook-5655c58dd6-dd6ts\" (UID: \"07446e4f-68a2-422e-b303-e3637d956a4c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.544538 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-v56pb" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.562371 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6wwq\" (UniqueName: \"kubernetes.io/projected/07446e4f-68a2-422e-b303-e3637d956a4c-kube-api-access-p6wwq\") pod \"cert-manager-webhook-5655c58dd6-dd6ts\" (UID: \"07446e4f-68a2-422e-b303-e3637d956a4c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.565922 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.794357 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-77j9j"] Oct 07 21:05:10 crc kubenswrapper[4926]: W1007 21:05:10.847525 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aa5ce8c_a725_4bd2_a504_887c768a64b6.slice/crio-f2f7a5902865a0393ca21e69c2c780e1f52f4212f5687fe06148846487586d92 WatchSource:0}: Error finding container f2f7a5902865a0393ca21e69c2c780e1f52f4212f5687fe06148846487586d92: Status 404 returned error can't find the container with id f2f7a5902865a0393ca21e69c2c780e1f52f4212f5687fe06148846487586d92 Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.852574 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.879590 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dd6ts"] Oct 07 21:05:10 crc kubenswrapper[4926]: I1007 21:05:10.887065 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-v56pb"] Oct 07 21:05:11 crc kubenswrapper[4926]: I1007 21:05:11.025758 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" event={"ID":"07446e4f-68a2-422e-b303-e3637d956a4c","Type":"ContainerStarted","Data":"b52273f7c6d63f2473c27e9f52be70d720118bcc04dffe9cc5686d1658c17ab5"} Oct 07 21:05:11 crc kubenswrapper[4926]: I1007 21:05:11.027567 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" event={"ID":"3aa5ce8c-a725-4bd2-a504-887c768a64b6","Type":"ContainerStarted","Data":"f2f7a5902865a0393ca21e69c2c780e1f52f4212f5687fe06148846487586d92"} Oct 07 21:05:11 crc kubenswrapper[4926]: I1007 21:05:11.032893 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-v56pb" event={"ID":"742ac735-fc39-4ab6-a072-834db359b8d8","Type":"ContainerStarted","Data":"db8767869f1bf5356bea7d96ea3b9bfadfbb95027a66ac6bf9a37d9021743cd2"} Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.063824 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" event={"ID":"07446e4f-68a2-422e-b303-e3637d956a4c","Type":"ContainerStarted","Data":"0084a9131a30e47b737f8114a08a4e4a500cd3f49af1155efa7abf805e47f6d8"} Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.064785 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.065900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" event={"ID":"3aa5ce8c-a725-4bd2-a504-887c768a64b6","Type":"ContainerStarted","Data":"39b45286720cd932402a3abb72910ba95b77f73aa23d6c32b0563d528860d698"} Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.067344 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-v56pb" event={"ID":"742ac735-fc39-4ab6-a072-834db359b8d8","Type":"ContainerStarted","Data":"d11d2a0324f2c8e014013419175390c61946b3b6ffabfdc46d5a7b940c70f31f"} Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.083758 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" podStartSLOduration=2.188566731 podStartE2EDuration="6.08372855s" podCreationTimestamp="2025-10-07 21:05:10 +0000 UTC" firstStartedPulling="2025-10-07 21:05:10.8935716 +0000 UTC m=+560.932152760" lastFinishedPulling="2025-10-07 21:05:14.788733389 +0000 UTC m=+564.827314579" observedRunningTime="2025-10-07 21:05:16.079765184 +0000 UTC m=+566.118346334" watchObservedRunningTime="2025-10-07 21:05:16.08372855 +0000 UTC m=+566.122309740" Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.143557 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-77j9j" podStartSLOduration=2.195464493 podStartE2EDuration="6.143534825s" podCreationTimestamp="2025-10-07 21:05:10 +0000 UTC" firstStartedPulling="2025-10-07 21:05:10.852232424 +0000 UTC m=+560.890813574" lastFinishedPulling="2025-10-07 21:05:14.800302736 +0000 UTC m=+564.838883906" observedRunningTime="2025-10-07 21:05:16.141769673 +0000 UTC m=+566.180350823" watchObservedRunningTime="2025-10-07 21:05:16.143534825 +0000 UTC m=+566.182115995" Oct 07 21:05:16 crc kubenswrapper[4926]: I1007 21:05:16.144778 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-v56pb" podStartSLOduration=2.170652229 podStartE2EDuration="6.144772621s" podCreationTimestamp="2025-10-07 21:05:10 +0000 UTC" firstStartedPulling="2025-10-07 21:05:10.891399647 +0000 UTC m=+560.929980797" lastFinishedPulling="2025-10-07 21:05:14.865520019 +0000 UTC m=+564.904101189" observedRunningTime="2025-10-07 21:05:16.110683936 +0000 UTC m=+566.149265096" watchObservedRunningTime="2025-10-07 21:05:16.144772621 +0000 UTC m=+566.183353771" Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.569952 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-dd6ts" Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.992981 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2zw6c"] Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.993643 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-controller" containerID="cri-o://f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" gracePeriod=30 Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.993666 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="sbdb" containerID="cri-o://ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" gracePeriod=30 Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.993787 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="nbdb" containerID="cri-o://9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" gracePeriod=30 Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.993857 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="northd" containerID="cri-o://77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" gracePeriod=30 Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.993931 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" gracePeriod=30 Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.993988 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-node" containerID="cri-o://ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" gracePeriod=30 Oct 07 21:05:20 crc kubenswrapper[4926]: I1007 21:05:20.994042 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-acl-logging" containerID="cri-o://e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" gracePeriod=30 Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.086848 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" containerID="cri-o://278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" gracePeriod=30 Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.403097 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/3.log" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.406656 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovn-acl-logging/0.log" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.407358 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovn-controller/0.log" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.407937 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478516 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6kptc"] Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478738 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-node" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478753 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-node" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478765 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="nbdb" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478774 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="nbdb" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478790 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478798 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478809 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-acl-logging" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478816 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-acl-logging" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478828 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="northd" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478835 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="northd" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478844 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kubecfg-setup" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478851 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kubecfg-setup" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478862 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478870 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478884 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478892 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478902 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="sbdb" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478909 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="sbdb" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478923 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478930 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478940 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478947 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.478958 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.478965 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479072 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="northd" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479088 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479098 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-node" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479108 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-acl-logging" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479119 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="sbdb" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479130 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovn-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479139 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="nbdb" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479151 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479160 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479169 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479179 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479189 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: E1007 21:05:21.479355 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.479366 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerName="ovnkube-controller" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.481453 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.490903 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-netd\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.490974 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491055 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-ovn-kubernetes\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491069 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491109 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-ovn\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491134 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491165 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-systemd\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491178 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491242 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-netns\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491303 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovn-node-metrics-cert\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491355 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-script-lib\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491425 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-var-lib-openvswitch\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491484 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-openvswitch\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491532 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-bin\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491575 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-node-log\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491634 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-etc-openvswitch\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491682 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-config\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491733 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-slash\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491766 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-log-socket\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491827 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-env-overrides\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491872 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-kubelet\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-systemd-units\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491957 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc8t4\" (UniqueName: \"kubernetes.io/projected/9b8c34cd-2651-4f25-8439-1897f0e937b4-kube-api-access-dc8t4\") pod \"9b8c34cd-2651-4f25-8439-1897f0e937b4\" (UID: \"9b8c34cd-2651-4f25-8439-1897f0e937b4\") " Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.491242 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492352 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-node-log" (OuterVolumeSpecName: "node-log") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492392 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492405 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492372 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492457 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492461 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492527 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-log-socket" (OuterVolumeSpecName: "log-socket") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492555 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492507 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-slash" (OuterVolumeSpecName: "host-slash") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492546 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.492878 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.493155 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.493268 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.493500 4926 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.493638 4926 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.493779 4926 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.493896 4926 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494007 4926 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494115 4926 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494254 4926 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-node-log\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494387 4926 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-slash\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494500 4926 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-log-socket\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494607 4926 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494716 4926 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.494826 4926 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.499216 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.499366 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b8c34cd-2651-4f25-8439-1897f0e937b4-kube-api-access-dc8t4" (OuterVolumeSpecName: "kube-api-access-dc8t4") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "kube-api-access-dc8t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.517872 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9b8c34cd-2651-4f25-8439-1897f0e937b4" (UID: "9b8c34cd-2651-4f25-8439-1897f0e937b4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.595767 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-systemd\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.595831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-run-netns\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.595866 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.595908 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-var-lib-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.595949 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh2sp\" (UniqueName: \"kubernetes.io/projected/332faa66-3eb5-442a-9557-0ba4214d2b9c-kube-api-access-dh2sp\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.595988 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-systemd-units\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596026 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-log-socket\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596065 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-etc-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596094 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-env-overrides\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596156 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovnkube-config\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596190 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-node-log\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596276 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596318 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovnkube-script-lib\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596355 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-cni-bin\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596391 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-ovn\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-cni-netd\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovn-node-metrics-cert\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596521 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-slash\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596551 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-kubelet\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596632 4926 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596660 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596686 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596710 4926 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596726 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596746 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b8c34cd-2651-4f25-8439-1897f0e937b4-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596762 4926 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b8c34cd-2651-4f25-8439-1897f0e937b4-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.596780 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc8t4\" (UniqueName: \"kubernetes.io/projected/9b8c34cd-2651-4f25-8439-1897f0e937b4-kube-api-access-dc8t4\") on node \"crc\" DevicePath \"\"" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.697955 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-var-lib-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698029 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh2sp\" (UniqueName: \"kubernetes.io/projected/332faa66-3eb5-442a-9557-0ba4214d2b9c-kube-api-access-dh2sp\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-systemd-units\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698104 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-log-socket\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698144 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-etc-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698175 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-env-overrides\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698259 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-etc-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698313 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-log-socket\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698166 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-var-lib-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698276 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovnkube-config\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698440 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-node-log\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698263 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-systemd-units\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698538 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698556 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-node-log\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698495 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovnkube-script-lib\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698673 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-cni-bin\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698721 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698777 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-ovn\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698808 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-cni-netd\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-slash\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698941 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovn-node-metrics-cert\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.698970 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-kubelet\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699054 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-systemd\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699089 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-run-netns\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699131 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699293 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-openvswitch\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-slash\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699373 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-cni-netd\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699391 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-env-overrides\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699466 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-cni-bin\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699486 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovnkube-config\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699537 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-kubelet\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699562 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-ovn\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699571 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-run-netns\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-run-systemd\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.699639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/332faa66-3eb5-442a-9557-0ba4214d2b9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.700361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovnkube-script-lib\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.706937 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/332faa66-3eb5-442a-9557-0ba4214d2b9c-ovn-node-metrics-cert\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.728153 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh2sp\" (UniqueName: \"kubernetes.io/projected/332faa66-3eb5-442a-9557-0ba4214d2b9c-kube-api-access-dh2sp\") pod \"ovnkube-node-6kptc\" (UID: \"332faa66-3eb5-442a-9557-0ba4214d2b9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: I1007 21:05:21.827013 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:21 crc kubenswrapper[4926]: W1007 21:05:21.854520 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod332faa66_3eb5_442a_9557_0ba4214d2b9c.slice/crio-1d3c77044dbc123946887e6ff4ba59d277fccc0e4d41504764cd93779fe6269f WatchSource:0}: Error finding container 1d3c77044dbc123946887e6ff4ba59d277fccc0e4d41504764cd93779fe6269f: Status 404 returned error can't find the container with id 1d3c77044dbc123946887e6ff4ba59d277fccc0e4d41504764cd93779fe6269f Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.128785 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovnkube-controller/3.log" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.132032 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovn-acl-logging/0.log" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.133098 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2zw6c_9b8c34cd-2651-4f25-8439-1897f0e937b4/ovn-controller/0.log" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.133954 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.133995 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134010 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134028 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134042 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134063 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134078 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" exitCode=143 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134096 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b8c34cd-2651-4f25-8439-1897f0e937b4" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" exitCode=143 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134266 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134291 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134313 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134332 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134369 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134369 4926 scope.go:117] "RemoveContainer" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134388 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134906 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134937 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134951 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134977 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.134989 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135000 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135011 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135040 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135073 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135089 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135100 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135111 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135121 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135131 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135142 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135152 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135166 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135179 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135228 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135253 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135271 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135285 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135299 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135313 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135327 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135341 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135355 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135372 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135383 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135399 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" event={"ID":"9b8c34cd-2651-4f25-8439-1897f0e937b4","Type":"ContainerDied","Data":"af39831457da4008b1b1a0ae8a2443ec8ddd724bbf7dd32621981b402e512d0b"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135417 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135430 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135440 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135451 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135462 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135472 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135483 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135495 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135505 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.135517 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.137338 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2zw6c" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.137792 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/2.log" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.138698 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/1.log" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.138759 4926 generic.go:334] "Generic (PLEG): container finished" podID="b1217995-75cc-4fc7-8f5f-7752d2e378e2" containerID="466b131acdb2f6ab6515194bccd251a8891942e98beea348991fa0cfa6ba158b" exitCode=2 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.138829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerDied","Data":"466b131acdb2f6ab6515194bccd251a8891942e98beea348991fa0cfa6ba158b"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.138854 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.139628 4926 scope.go:117] "RemoveContainer" containerID="466b131acdb2f6ab6515194bccd251a8891942e98beea348991fa0cfa6ba158b" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.140105 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2tlv9_openshift-multus(b1217995-75cc-4fc7-8f5f-7752d2e378e2)\"" pod="openshift-multus/multus-2tlv9" podUID="b1217995-75cc-4fc7-8f5f-7752d2e378e2" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.143541 4926 generic.go:334] "Generic (PLEG): container finished" podID="332faa66-3eb5-442a-9557-0ba4214d2b9c" containerID="6d520a510034e685aff970331420cc554fcf7cfa38872934ac32ab7683defe1d" exitCode=0 Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.143598 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerDied","Data":"6d520a510034e685aff970331420cc554fcf7cfa38872934ac32ab7683defe1d"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.143633 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"1d3c77044dbc123946887e6ff4ba59d277fccc0e4d41504764cd93779fe6269f"} Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.163949 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.209461 4926 scope.go:117] "RemoveContainer" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.242098 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2zw6c"] Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.252091 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2zw6c"] Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.279741 4926 scope.go:117] "RemoveContainer" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.307066 4926 scope.go:117] "RemoveContainer" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.337670 4926 scope.go:117] "RemoveContainer" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.359706 4926 scope.go:117] "RemoveContainer" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.387633 4926 scope.go:117] "RemoveContainer" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.440882 4926 scope.go:117] "RemoveContainer" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.465160 4926 scope.go:117] "RemoveContainer" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.502394 4926 scope.go:117] "RemoveContainer" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.502916 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": container with ID starting with 278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07 not found: ID does not exist" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.502968 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} err="failed to get container status \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": rpc error: code = NotFound desc = could not find container \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": container with ID starting with 278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.503009 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.503414 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": container with ID starting with a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918 not found: ID does not exist" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.503460 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} err="failed to get container status \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": rpc error: code = NotFound desc = could not find container \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": container with ID starting with a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.503489 4926 scope.go:117] "RemoveContainer" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.504110 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": container with ID starting with ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7 not found: ID does not exist" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.504159 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} err="failed to get container status \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": rpc error: code = NotFound desc = could not find container \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": container with ID starting with ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.504187 4926 scope.go:117] "RemoveContainer" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.504615 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": container with ID starting with 9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62 not found: ID does not exist" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.504656 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} err="failed to get container status \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": rpc error: code = NotFound desc = could not find container \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": container with ID starting with 9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.504683 4926 scope.go:117] "RemoveContainer" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.505108 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": container with ID starting with 77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8 not found: ID does not exist" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.505150 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} err="failed to get container status \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": rpc error: code = NotFound desc = could not find container \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": container with ID starting with 77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.505177 4926 scope.go:117] "RemoveContainer" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.505987 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": container with ID starting with 105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876 not found: ID does not exist" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.506031 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} err="failed to get container status \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": rpc error: code = NotFound desc = could not find container \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": container with ID starting with 105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.506060 4926 scope.go:117] "RemoveContainer" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.506455 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": container with ID starting with ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263 not found: ID does not exist" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.506497 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} err="failed to get container status \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": rpc error: code = NotFound desc = could not find container \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": container with ID starting with ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.506523 4926 scope.go:117] "RemoveContainer" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.507111 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": container with ID starting with e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012 not found: ID does not exist" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.507153 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} err="failed to get container status \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": rpc error: code = NotFound desc = could not find container \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": container with ID starting with e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.507181 4926 scope.go:117] "RemoveContainer" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.507742 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": container with ID starting with f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808 not found: ID does not exist" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.507793 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} err="failed to get container status \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": rpc error: code = NotFound desc = could not find container \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": container with ID starting with f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.507822 4926 scope.go:117] "RemoveContainer" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" Oct 07 21:05:22 crc kubenswrapper[4926]: E1007 21:05:22.508097 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": container with ID starting with aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3 not found: ID does not exist" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.508141 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} err="failed to get container status \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": rpc error: code = NotFound desc = could not find container \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": container with ID starting with aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.508167 4926 scope.go:117] "RemoveContainer" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.508550 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} err="failed to get container status \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": rpc error: code = NotFound desc = could not find container \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": container with ID starting with 278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.508588 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.509646 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} err="failed to get container status \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": rpc error: code = NotFound desc = could not find container \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": container with ID starting with a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.509687 4926 scope.go:117] "RemoveContainer" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.510293 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} err="failed to get container status \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": rpc error: code = NotFound desc = could not find container \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": container with ID starting with ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.510381 4926 scope.go:117] "RemoveContainer" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.510795 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} err="failed to get container status \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": rpc error: code = NotFound desc = could not find container \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": container with ID starting with 9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.510877 4926 scope.go:117] "RemoveContainer" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.511407 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} err="failed to get container status \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": rpc error: code = NotFound desc = could not find container \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": container with ID starting with 77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.511491 4926 scope.go:117] "RemoveContainer" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.512057 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} err="failed to get container status \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": rpc error: code = NotFound desc = could not find container \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": container with ID starting with 105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.512099 4926 scope.go:117] "RemoveContainer" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.512472 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} err="failed to get container status \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": rpc error: code = NotFound desc = could not find container \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": container with ID starting with ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.512542 4926 scope.go:117] "RemoveContainer" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.512957 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} err="failed to get container status \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": rpc error: code = NotFound desc = could not find container \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": container with ID starting with e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.513031 4926 scope.go:117] "RemoveContainer" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.513517 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} err="failed to get container status \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": rpc error: code = NotFound desc = could not find container \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": container with ID starting with f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.513594 4926 scope.go:117] "RemoveContainer" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.513886 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} err="failed to get container status \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": rpc error: code = NotFound desc = could not find container \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": container with ID starting with aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.513965 4926 scope.go:117] "RemoveContainer" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.514270 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} err="failed to get container status \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": rpc error: code = NotFound desc = could not find container \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": container with ID starting with 278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.514310 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.514528 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} err="failed to get container status \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": rpc error: code = NotFound desc = could not find container \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": container with ID starting with a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.514593 4926 scope.go:117] "RemoveContainer" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.514855 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} err="failed to get container status \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": rpc error: code = NotFound desc = could not find container \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": container with ID starting with ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.514929 4926 scope.go:117] "RemoveContainer" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.515507 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} err="failed to get container status \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": rpc error: code = NotFound desc = could not find container \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": container with ID starting with 9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.515581 4926 scope.go:117] "RemoveContainer" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.515975 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} err="failed to get container status \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": rpc error: code = NotFound desc = could not find container \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": container with ID starting with 77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.516061 4926 scope.go:117] "RemoveContainer" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.516596 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} err="failed to get container status \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": rpc error: code = NotFound desc = could not find container \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": container with ID starting with 105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.516634 4926 scope.go:117] "RemoveContainer" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.516933 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} err="failed to get container status \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": rpc error: code = NotFound desc = could not find container \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": container with ID starting with ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.516966 4926 scope.go:117] "RemoveContainer" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.517871 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} err="failed to get container status \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": rpc error: code = NotFound desc = could not find container \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": container with ID starting with e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.517945 4926 scope.go:117] "RemoveContainer" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.518657 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} err="failed to get container status \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": rpc error: code = NotFound desc = could not find container \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": container with ID starting with f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.518701 4926 scope.go:117] "RemoveContainer" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.519313 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} err="failed to get container status \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": rpc error: code = NotFound desc = could not find container \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": container with ID starting with aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.519354 4926 scope.go:117] "RemoveContainer" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.519779 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} err="failed to get container status \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": rpc error: code = NotFound desc = could not find container \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": container with ID starting with 278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.519824 4926 scope.go:117] "RemoveContainer" containerID="a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.520073 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918"} err="failed to get container status \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": rpc error: code = NotFound desc = could not find container \"a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918\": container with ID starting with a5aced1d1df3ed88e7b35ab959e1254cc2e767fb6d1150e6c4249b16d2cc4918 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.520108 4926 scope.go:117] "RemoveContainer" containerID="ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.520340 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7"} err="failed to get container status \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": rpc error: code = NotFound desc = could not find container \"ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7\": container with ID starting with ae5d92ac4999dff6bed8858e74e5df7f1a52517547414c5be2b75bb07bee1bf7 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.520375 4926 scope.go:117] "RemoveContainer" containerID="9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.522652 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62"} err="failed to get container status \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": rpc error: code = NotFound desc = could not find container \"9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62\": container with ID starting with 9fe85bd345c68175cdc7d31a605db16dd02f6cd76cf68f067b114026701e0f62 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.522702 4926 scope.go:117] "RemoveContainer" containerID="77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.523002 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8"} err="failed to get container status \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": rpc error: code = NotFound desc = could not find container \"77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8\": container with ID starting with 77c8c41a98226a7241e0a48fc4e5566133c793db46babafd7899644f14cb53d8 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.523043 4926 scope.go:117] "RemoveContainer" containerID="105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.523551 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876"} err="failed to get container status \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": rpc error: code = NotFound desc = could not find container \"105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876\": container with ID starting with 105362341a5559792e3187389dea24c13df6438b413ac3a39ea573eac2c5e876 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.523635 4926 scope.go:117] "RemoveContainer" containerID="ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.524008 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263"} err="failed to get container status \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": rpc error: code = NotFound desc = could not find container \"ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263\": container with ID starting with ce91be3d059fa09a495a9d696146d146e83aa22319b3c6290be27deecb43f263 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.524046 4926 scope.go:117] "RemoveContainer" containerID="e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.524319 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012"} err="failed to get container status \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": rpc error: code = NotFound desc = could not find container \"e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012\": container with ID starting with e47f223fbaf88e8668555ca33c5705484a8fa61828e9bc976a611cf5fd14f012 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.524409 4926 scope.go:117] "RemoveContainer" containerID="f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.524773 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808"} err="failed to get container status \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": rpc error: code = NotFound desc = could not find container \"f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808\": container with ID starting with f14df7339aceeedcb7fd62347a5580d29a0b02842bbb8161a68ae1ffd37a2808 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.524867 4926 scope.go:117] "RemoveContainer" containerID="aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.525318 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3"} err="failed to get container status \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": rpc error: code = NotFound desc = could not find container \"aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3\": container with ID starting with aa47dd67f071ffbb83c18b0876f7d49de9c96d432f0c2349504aa6568d10abe3 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.525396 4926 scope.go:117] "RemoveContainer" containerID="278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.525675 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07"} err="failed to get container status \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": rpc error: code = NotFound desc = could not find container \"278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07\": container with ID starting with 278f572473b06014f37db83c505b98143ab4e6f292faad2978e2232d38173b07 not found: ID does not exist" Oct 07 21:05:22 crc kubenswrapper[4926]: I1007 21:05:22.688037 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b8c34cd-2651-4f25-8439-1897f0e937b4" path="/var/lib/kubelet/pods/9b8c34cd-2651-4f25-8439-1897f0e937b4/volumes" Oct 07 21:05:23 crc kubenswrapper[4926]: I1007 21:05:23.166310 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"b67fbd15f37af79caf77bd18d16a2d39a53d783effae478c0443d13bca641585"} Oct 07 21:05:23 crc kubenswrapper[4926]: I1007 21:05:23.166360 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"470683c7ef285a755a15a9c05d6f2746fa77e764a023733a295f1f758813b81f"} Oct 07 21:05:23 crc kubenswrapper[4926]: I1007 21:05:23.166378 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"bcbad6b545660a32afe5fec9ef0dcdcff030b5c1f73fa8020f08d15a0d3c50c9"} Oct 07 21:05:23 crc kubenswrapper[4926]: I1007 21:05:23.166391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"fe0afe80ff500d8040f34783e4ba5e8fca6243cba1afe76573eb1f09e6fa40cd"} Oct 07 21:05:23 crc kubenswrapper[4926]: I1007 21:05:23.166403 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"ea26b66166fa7c6d5e82502f0bbbc3f7c962805f952341dff2f0bcad6c556178"} Oct 07 21:05:23 crc kubenswrapper[4926]: I1007 21:05:23.166414 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"976769cde6ce003e8a1b13a494c4e75c0462416fcebf33ddc64b87f2548e62b4"} Oct 07 21:05:26 crc kubenswrapper[4926]: I1007 21:05:26.190502 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"58283d8cb7fdbcae52da89f671418298b9e3695e50676517dc192b756c76cba9"} Oct 07 21:05:28 crc kubenswrapper[4926]: I1007 21:05:28.218145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" event={"ID":"332faa66-3eb5-442a-9557-0ba4214d2b9c","Type":"ContainerStarted","Data":"06a0c49c91dd70b2261202504a12863fee357abf4d4f6ce55d7eec4a40cb8002"} Oct 07 21:05:28 crc kubenswrapper[4926]: I1007 21:05:28.218660 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:28 crc kubenswrapper[4926]: I1007 21:05:28.218676 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:28 crc kubenswrapper[4926]: I1007 21:05:28.256620 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:28 crc kubenswrapper[4926]: I1007 21:05:28.280472 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" podStartSLOduration=7.280382519 podStartE2EDuration="7.280382519s" podCreationTimestamp="2025-10-07 21:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:05:28.266116833 +0000 UTC m=+578.304698053" watchObservedRunningTime="2025-10-07 21:05:28.280382519 +0000 UTC m=+578.318963739" Oct 07 21:05:29 crc kubenswrapper[4926]: I1007 21:05:29.225467 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:29 crc kubenswrapper[4926]: I1007 21:05:29.297546 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:05:33 crc kubenswrapper[4926]: I1007 21:05:33.209434 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:05:33 crc kubenswrapper[4926]: I1007 21:05:33.211531 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:05:33 crc kubenswrapper[4926]: I1007 21:05:33.211647 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:05:33 crc kubenswrapper[4926]: I1007 21:05:33.212750 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea4910515b9deac1bdd55447e1b44ec9155b5c86397c1d6d38253d2a05662255"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:05:33 crc kubenswrapper[4926]: I1007 21:05:33.212873 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://ea4910515b9deac1bdd55447e1b44ec9155b5c86397c1d6d38253d2a05662255" gracePeriod=600 Oct 07 21:05:34 crc kubenswrapper[4926]: I1007 21:05:34.262609 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="ea4910515b9deac1bdd55447e1b44ec9155b5c86397c1d6d38253d2a05662255" exitCode=0 Oct 07 21:05:34 crc kubenswrapper[4926]: I1007 21:05:34.262720 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"ea4910515b9deac1bdd55447e1b44ec9155b5c86397c1d6d38253d2a05662255"} Oct 07 21:05:34 crc kubenswrapper[4926]: I1007 21:05:34.263147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"b152fe038f21276c6a7128f942977a43dfd86caa4be6d3a00a66f622f88a2161"} Oct 07 21:05:34 crc kubenswrapper[4926]: I1007 21:05:34.263179 4926 scope.go:117] "RemoveContainer" containerID="03c2adc15ef42fc76b9b086cb5172f247b75e21cd7f7824690f65b2c4a8e8842" Oct 07 21:05:35 crc kubenswrapper[4926]: I1007 21:05:35.679245 4926 scope.go:117] "RemoveContainer" containerID="466b131acdb2f6ab6515194bccd251a8891942e98beea348991fa0cfa6ba158b" Oct 07 21:05:35 crc kubenswrapper[4926]: E1007 21:05:35.680048 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2tlv9_openshift-multus(b1217995-75cc-4fc7-8f5f-7752d2e378e2)\"" pod="openshift-multus/multus-2tlv9" podUID="b1217995-75cc-4fc7-8f5f-7752d2e378e2" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.060515 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf"] Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.066559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.070463 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf"] Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.070692 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.131356 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.131707 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.131968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgkzb\" (UniqueName: \"kubernetes.io/projected/b7515481-e886-419f-a65b-ca89f922ef2e-kube-api-access-pgkzb\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.233980 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.234115 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgkzb\" (UniqueName: \"kubernetes.io/projected/b7515481-e886-419f-a65b-ca89f922ef2e-kube-api-access-pgkzb\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.234246 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.235378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.235498 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.267760 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgkzb\" (UniqueName: \"kubernetes.io/projected/b7515481-e886-419f-a65b-ca89f922ef2e-kube-api-access-pgkzb\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: I1007 21:05:49.395286 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: E1007 21:05:49.438687 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(70d8085886fd05dc2eb8abf50747a076270ca38033e1318f841b61d7c0315452): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 21:05:49 crc kubenswrapper[4926]: E1007 21:05:49.438779 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(70d8085886fd05dc2eb8abf50747a076270ca38033e1318f841b61d7c0315452): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: E1007 21:05:49.438820 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(70d8085886fd05dc2eb8abf50747a076270ca38033e1318f841b61d7c0315452): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:49 crc kubenswrapper[4926]: E1007 21:05:49.438901 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace(b7515481-e886-419f-a65b-ca89f922ef2e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace(b7515481-e886-419f-a65b-ca89f922ef2e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(70d8085886fd05dc2eb8abf50747a076270ca38033e1318f841b61d7c0315452): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" Oct 07 21:05:50 crc kubenswrapper[4926]: I1007 21:05:50.379148 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:50 crc kubenswrapper[4926]: I1007 21:05:50.381449 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:50 crc kubenswrapper[4926]: E1007 21:05:50.417067 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(c212984716a53097fd2c0da9b0f1615858f91e0155efaa9c72cf19e78bbf1fb7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 21:05:50 crc kubenswrapper[4926]: E1007 21:05:50.417151 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(c212984716a53097fd2c0da9b0f1615858f91e0155efaa9c72cf19e78bbf1fb7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:50 crc kubenswrapper[4926]: E1007 21:05:50.417189 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(c212984716a53097fd2c0da9b0f1615858f91e0155efaa9c72cf19e78bbf1fb7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:05:50 crc kubenswrapper[4926]: E1007 21:05:50.417288 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace(b7515481-e886-419f-a65b-ca89f922ef2e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace(b7515481-e886-419f-a65b-ca89f922ef2e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_openshift-marketplace_b7515481-e886-419f-a65b-ca89f922ef2e_0(c212984716a53097fd2c0da9b0f1615858f91e0155efaa9c72cf19e78bbf1fb7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" Oct 07 21:05:50 crc kubenswrapper[4926]: I1007 21:05:50.684912 4926 scope.go:117] "RemoveContainer" containerID="466b131acdb2f6ab6515194bccd251a8891942e98beea348991fa0cfa6ba158b" Oct 07 21:05:50 crc kubenswrapper[4926]: I1007 21:05:50.879374 4926 scope.go:117] "RemoveContainer" containerID="31d4332851c42cb0f64410af24ee4d214f36a2edd98fef2a51d4ab67717ed4b5" Oct 07 21:05:51 crc kubenswrapper[4926]: I1007 21:05:51.391592 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2tlv9_b1217995-75cc-4fc7-8f5f-7752d2e378e2/kube-multus/2.log" Oct 07 21:05:51 crc kubenswrapper[4926]: I1007 21:05:51.391690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2tlv9" event={"ID":"b1217995-75cc-4fc7-8f5f-7752d2e378e2","Type":"ContainerStarted","Data":"556a8b1795dc94ff43dfb50becf760cc17417dfb62fce9570d340368e8b84412"} Oct 07 21:05:51 crc kubenswrapper[4926]: I1007 21:05:51.863060 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6kptc" Oct 07 21:06:05 crc kubenswrapper[4926]: I1007 21:06:05.678475 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:06:05 crc kubenswrapper[4926]: I1007 21:06:05.679566 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:06:05 crc kubenswrapper[4926]: I1007 21:06:05.947120 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf"] Oct 07 21:06:05 crc kubenswrapper[4926]: W1007 21:06:05.967397 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7515481_e886_419f_a65b_ca89f922ef2e.slice/crio-d0d20daaccf5be4aa041e3a7ef59a6a176cca262f0d0fc4b2ffcffd05a81dcf4 WatchSource:0}: Error finding container d0d20daaccf5be4aa041e3a7ef59a6a176cca262f0d0fc4b2ffcffd05a81dcf4: Status 404 returned error can't find the container with id d0d20daaccf5be4aa041e3a7ef59a6a176cca262f0d0fc4b2ffcffd05a81dcf4 Oct 07 21:06:06 crc kubenswrapper[4926]: I1007 21:06:06.490029 4926 generic.go:334] "Generic (PLEG): container finished" podID="b7515481-e886-419f-a65b-ca89f922ef2e" containerID="e84a2075f138f170207920f6627943321e08480a9b11d056dea4a92a1107a8ce" exitCode=0 Oct 07 21:06:06 crc kubenswrapper[4926]: I1007 21:06:06.490101 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" event={"ID":"b7515481-e886-419f-a65b-ca89f922ef2e","Type":"ContainerDied","Data":"e84a2075f138f170207920f6627943321e08480a9b11d056dea4a92a1107a8ce"} Oct 07 21:06:06 crc kubenswrapper[4926]: I1007 21:06:06.490147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" event={"ID":"b7515481-e886-419f-a65b-ca89f922ef2e","Type":"ContainerStarted","Data":"d0d20daaccf5be4aa041e3a7ef59a6a176cca262f0d0fc4b2ffcffd05a81dcf4"} Oct 07 21:06:08 crc kubenswrapper[4926]: I1007 21:06:08.506790 4926 generic.go:334] "Generic (PLEG): container finished" podID="b7515481-e886-419f-a65b-ca89f922ef2e" containerID="bdc76eb915a661314c125a23c7ab8290c32de647b7cd594dfbfa645a66b8ba21" exitCode=0 Oct 07 21:06:08 crc kubenswrapper[4926]: I1007 21:06:08.506875 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" event={"ID":"b7515481-e886-419f-a65b-ca89f922ef2e","Type":"ContainerDied","Data":"bdc76eb915a661314c125a23c7ab8290c32de647b7cd594dfbfa645a66b8ba21"} Oct 07 21:06:09 crc kubenswrapper[4926]: I1007 21:06:09.529050 4926 generic.go:334] "Generic (PLEG): container finished" podID="b7515481-e886-419f-a65b-ca89f922ef2e" containerID="9db552a08f14ef6b297e46e265fe7758aa6c04e7f74460c067a1f9e7754741b3" exitCode=0 Oct 07 21:06:09 crc kubenswrapper[4926]: I1007 21:06:09.529143 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" event={"ID":"b7515481-e886-419f-a65b-ca89f922ef2e","Type":"ContainerDied","Data":"9db552a08f14ef6b297e46e265fe7758aa6c04e7f74460c067a1f9e7754741b3"} Oct 07 21:06:10 crc kubenswrapper[4926]: I1007 21:06:10.871721 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:06:10 crc kubenswrapper[4926]: I1007 21:06:10.973635 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-util\") pod \"b7515481-e886-419f-a65b-ca89f922ef2e\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " Oct 07 21:06:10 crc kubenswrapper[4926]: I1007 21:06:10.973751 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgkzb\" (UniqueName: \"kubernetes.io/projected/b7515481-e886-419f-a65b-ca89f922ef2e-kube-api-access-pgkzb\") pod \"b7515481-e886-419f-a65b-ca89f922ef2e\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " Oct 07 21:06:10 crc kubenswrapper[4926]: I1007 21:06:10.973821 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-bundle\") pod \"b7515481-e886-419f-a65b-ca89f922ef2e\" (UID: \"b7515481-e886-419f-a65b-ca89f922ef2e\") " Oct 07 21:06:10 crc kubenswrapper[4926]: I1007 21:06:10.981503 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-bundle" (OuterVolumeSpecName: "bundle") pod "b7515481-e886-419f-a65b-ca89f922ef2e" (UID: "b7515481-e886-419f-a65b-ca89f922ef2e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:06:10 crc kubenswrapper[4926]: I1007 21:06:10.983108 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7515481-e886-419f-a65b-ca89f922ef2e-kube-api-access-pgkzb" (OuterVolumeSpecName: "kube-api-access-pgkzb") pod "b7515481-e886-419f-a65b-ca89f922ef2e" (UID: "b7515481-e886-419f-a65b-ca89f922ef2e"). InnerVolumeSpecName "kube-api-access-pgkzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.016318 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-util" (OuterVolumeSpecName: "util") pod "b7515481-e886-419f-a65b-ca89f922ef2e" (UID: "b7515481-e886-419f-a65b-ca89f922ef2e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.076052 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-util\") on node \"crc\" DevicePath \"\"" Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.076111 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgkzb\" (UniqueName: \"kubernetes.io/projected/b7515481-e886-419f-a65b-ca89f922ef2e-kube-api-access-pgkzb\") on node \"crc\" DevicePath \"\"" Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.076137 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7515481-e886-419f-a65b-ca89f922ef2e-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.548402 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" event={"ID":"b7515481-e886-419f-a65b-ca89f922ef2e","Type":"ContainerDied","Data":"d0d20daaccf5be4aa041e3a7ef59a6a176cca262f0d0fc4b2ffcffd05a81dcf4"} Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.548475 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0d20daaccf5be4aa041e3a7ef59a6a176cca262f0d0fc4b2ffcffd05a81dcf4" Oct 07 21:06:11 crc kubenswrapper[4926]: I1007 21:06:11.548547 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.591447 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9"] Oct 07 21:06:21 crc kubenswrapper[4926]: E1007 21:06:21.592136 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="extract" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.592148 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="extract" Oct 07 21:06:21 crc kubenswrapper[4926]: E1007 21:06:21.592162 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="pull" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.592169 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="pull" Oct 07 21:06:21 crc kubenswrapper[4926]: E1007 21:06:21.592181 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="util" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.592187 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="util" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.592313 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7515481-e886-419f-a65b-ca89f922ef2e" containerName="extract" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.592690 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.594615 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.594642 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-lhght" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.594786 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.602077 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.708376 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.709042 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.711526 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-hrsvz" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.711869 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.719709 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.720699 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.730245 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.734211 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.771530 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qb2f\" (UniqueName: \"kubernetes.io/projected/0847160c-b3dd-4539-a141-d263faa41fd8-kube-api-access-4qb2f\") pod \"obo-prometheus-operator-7c8cf85677-hg4l9\" (UID: \"0847160c-b3dd-4539-a141-d263faa41fd8\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.831753 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-9pr8c"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.832676 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.836631 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.838624 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-bl7lm" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.846612 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-9pr8c"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.873713 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qb2f\" (UniqueName: \"kubernetes.io/projected/0847160c-b3dd-4539-a141-d263faa41fd8-kube-api-access-4qb2f\") pod \"obo-prometheus-operator-7c8cf85677-hg4l9\" (UID: \"0847160c-b3dd-4539-a141-d263faa41fd8\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.873829 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f92dd43-5e91-41a3-bd84-8de059b4e9ee-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs\" (UID: \"9f92dd43-5e91-41a3-bd84-8de059b4e9ee\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.873865 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/22bad45c-639a-4d86-aa08-1dd5f9b8ec5d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98\" (UID: \"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.873899 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f92dd43-5e91-41a3-bd84-8de059b4e9ee-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs\" (UID: \"9f92dd43-5e91-41a3-bd84-8de059b4e9ee\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.873928 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/22bad45c-639a-4d86-aa08-1dd5f9b8ec5d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98\" (UID: \"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.900770 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qb2f\" (UniqueName: \"kubernetes.io/projected/0847160c-b3dd-4539-a141-d263faa41fd8-kube-api-access-4qb2f\") pod \"obo-prometheus-operator-7c8cf85677-hg4l9\" (UID: \"0847160c-b3dd-4539-a141-d263faa41fd8\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.907320 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.936540 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-9vvdj"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.937273 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.939232 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-l9bpm" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.956542 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-9vvdj"] Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.974901 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f92dd43-5e91-41a3-bd84-8de059b4e9ee-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs\" (UID: \"9f92dd43-5e91-41a3-bd84-8de059b4e9ee\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.975165 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/22bad45c-639a-4d86-aa08-1dd5f9b8ec5d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98\" (UID: \"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.975313 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f92dd43-5e91-41a3-bd84-8de059b4e9ee-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs\" (UID: \"9f92dd43-5e91-41a3-bd84-8de059b4e9ee\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.975424 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/22bad45c-639a-4d86-aa08-1dd5f9b8ec5d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98\" (UID: \"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.975535 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d83bdfa-2009-43f5-93b1-cdc8e568020d-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-9pr8c\" (UID: \"4d83bdfa-2009-43f5-93b1-cdc8e568020d\") " pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.975637 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p26jw\" (UniqueName: \"kubernetes.io/projected/4d83bdfa-2009-43f5-93b1-cdc8e568020d-kube-api-access-p26jw\") pod \"observability-operator-cc5f78dfc-9pr8c\" (UID: \"4d83bdfa-2009-43f5-93b1-cdc8e568020d\") " pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.981066 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f92dd43-5e91-41a3-bd84-8de059b4e9ee-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs\" (UID: \"9f92dd43-5e91-41a3-bd84-8de059b4e9ee\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.981066 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/22bad45c-639a-4d86-aa08-1dd5f9b8ec5d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98\" (UID: \"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.981927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f92dd43-5e91-41a3-bd84-8de059b4e9ee-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs\" (UID: \"9f92dd43-5e91-41a3-bd84-8de059b4e9ee\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:21 crc kubenswrapper[4926]: I1007 21:06:21.986636 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/22bad45c-639a-4d86-aa08-1dd5f9b8ec5d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98\" (UID: \"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.023952 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.035473 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.076948 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r64p\" (UniqueName: \"kubernetes.io/projected/b4f527c2-fe8d-4da9-8477-2f02507a6178-kube-api-access-7r64p\") pod \"perses-operator-54bc95c9fb-9vvdj\" (UID: \"b4f527c2-fe8d-4da9-8477-2f02507a6178\") " pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.077024 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d83bdfa-2009-43f5-93b1-cdc8e568020d-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-9pr8c\" (UID: \"4d83bdfa-2009-43f5-93b1-cdc8e568020d\") " pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.077157 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p26jw\" (UniqueName: \"kubernetes.io/projected/4d83bdfa-2009-43f5-93b1-cdc8e568020d-kube-api-access-p26jw\") pod \"observability-operator-cc5f78dfc-9pr8c\" (UID: \"4d83bdfa-2009-43f5-93b1-cdc8e568020d\") " pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.077333 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b4f527c2-fe8d-4da9-8477-2f02507a6178-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-9vvdj\" (UID: \"b4f527c2-fe8d-4da9-8477-2f02507a6178\") " pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.082083 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d83bdfa-2009-43f5-93b1-cdc8e568020d-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-9pr8c\" (UID: \"4d83bdfa-2009-43f5-93b1-cdc8e568020d\") " pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.093373 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p26jw\" (UniqueName: \"kubernetes.io/projected/4d83bdfa-2009-43f5-93b1-cdc8e568020d-kube-api-access-p26jw\") pod \"observability-operator-cc5f78dfc-9pr8c\" (UID: \"4d83bdfa-2009-43f5-93b1-cdc8e568020d\") " pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.146061 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9"] Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.146350 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.183862 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b4f527c2-fe8d-4da9-8477-2f02507a6178-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-9vvdj\" (UID: \"b4f527c2-fe8d-4da9-8477-2f02507a6178\") " pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.183912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r64p\" (UniqueName: \"kubernetes.io/projected/b4f527c2-fe8d-4da9-8477-2f02507a6178-kube-api-access-7r64p\") pod \"perses-operator-54bc95c9fb-9vvdj\" (UID: \"b4f527c2-fe8d-4da9-8477-2f02507a6178\") " pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.184645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b4f527c2-fe8d-4da9-8477-2f02507a6178-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-9vvdj\" (UID: \"b4f527c2-fe8d-4da9-8477-2f02507a6178\") " pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.219408 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r64p\" (UniqueName: \"kubernetes.io/projected/b4f527c2-fe8d-4da9-8477-2f02507a6178-kube-api-access-7r64p\") pod \"perses-operator-54bc95c9fb-9vvdj\" (UID: \"b4f527c2-fe8d-4da9-8477-2f02507a6178\") " pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.264508 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.332166 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs"] Oct 07 21:06:22 crc kubenswrapper[4926]: W1007 21:06:22.349988 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f92dd43_5e91_41a3_bd84_8de059b4e9ee.slice/crio-be459e7a4f1bacc1357887e896f8a65437fa621cc4deaa373906375e9bdb89d9 WatchSource:0}: Error finding container be459e7a4f1bacc1357887e896f8a65437fa621cc4deaa373906375e9bdb89d9: Status 404 returned error can't find the container with id be459e7a4f1bacc1357887e896f8a65437fa621cc4deaa373906375e9bdb89d9 Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.488720 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-9pr8c"] Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.535637 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-9vvdj"] Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.607217 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" event={"ID":"0847160c-b3dd-4539-a141-d263faa41fd8","Type":"ContainerStarted","Data":"e8a71dcdcfd21354d6cb27e2a05e9a962f9bfd4a81abb428d75f72c4185126c1"} Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.608293 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" event={"ID":"b4f527c2-fe8d-4da9-8477-2f02507a6178","Type":"ContainerStarted","Data":"e27b6cd3f66cbfe78380f473cb94e123dae579a954686f993bd1eed2e36f9100"} Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.611486 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" event={"ID":"4d83bdfa-2009-43f5-93b1-cdc8e568020d","Type":"ContainerStarted","Data":"4737b6cfbd6a6cc99eb8f060c18aa7ce3aac79b572a4f9836663438b7c450090"} Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.612557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" event={"ID":"9f92dd43-5e91-41a3-bd84-8de059b4e9ee","Type":"ContainerStarted","Data":"be459e7a4f1bacc1357887e896f8a65437fa621cc4deaa373906375e9bdb89d9"} Oct 07 21:06:22 crc kubenswrapper[4926]: I1007 21:06:22.628174 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98"] Oct 07 21:06:23 crc kubenswrapper[4926]: I1007 21:06:23.619885 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" event={"ID":"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d","Type":"ContainerStarted","Data":"080a0fbc929a72fa13f9e46add5cbd3850cbe819c64f7107aeb3e9e7324abb8d"} Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.738820 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" event={"ID":"9f92dd43-5e91-41a3-bd84-8de059b4e9ee","Type":"ContainerStarted","Data":"348c7729dbb95def6c11c4f050ff29d1c6d2e33663e601d588dac5c65e83d8bc"} Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.740380 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" event={"ID":"22bad45c-639a-4d86-aa08-1dd5f9b8ec5d","Type":"ContainerStarted","Data":"443440adafe769c031de5bfdbb6d96a475fd34bc055d9aa0da2d15dd4dcc8016"} Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.741776 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" event={"ID":"0847160c-b3dd-4539-a141-d263faa41fd8","Type":"ContainerStarted","Data":"b0a1e03fde608cdbadfa2e8fd7e4648b50880bdd49415963d62d8fed0a86a7c9"} Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.743284 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" event={"ID":"b4f527c2-fe8d-4da9-8477-2f02507a6178","Type":"ContainerStarted","Data":"7a76a90b65879d00342798e40368a550c4515e3aa2cad124e2a52cc753df01ab"} Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.743384 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.746115 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" event={"ID":"4d83bdfa-2009-43f5-93b1-cdc8e568020d","Type":"ContainerStarted","Data":"6ece225a381d0c6b87426a5e944bef1b5657dcbe5d43e80d8017ba0420cab052"} Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.746358 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.748444 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.766012 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs" podStartSLOduration=1.8626447069999998 podStartE2EDuration="16.765997801s" podCreationTimestamp="2025-10-07 21:06:21 +0000 UTC" firstStartedPulling="2025-10-07 21:06:22.356177022 +0000 UTC m=+632.394758172" lastFinishedPulling="2025-10-07 21:06:37.259530116 +0000 UTC m=+647.298111266" observedRunningTime="2025-10-07 21:06:37.765297121 +0000 UTC m=+647.803878281" watchObservedRunningTime="2025-10-07 21:06:37.765997801 +0000 UTC m=+647.804578951" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.814962 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-hg4l9" podStartSLOduration=1.671896636 podStartE2EDuration="16.814946101s" podCreationTimestamp="2025-10-07 21:06:21 +0000 UTC" firstStartedPulling="2025-10-07 21:06:22.206337881 +0000 UTC m=+632.244919031" lastFinishedPulling="2025-10-07 21:06:37.349387346 +0000 UTC m=+647.387968496" observedRunningTime="2025-10-07 21:06:37.812978265 +0000 UTC m=+647.851559415" watchObservedRunningTime="2025-10-07 21:06:37.814946101 +0000 UTC m=+647.853527251" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.837456 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" podStartSLOduration=2.129076379 podStartE2EDuration="16.837439294s" podCreationTimestamp="2025-10-07 21:06:21 +0000 UTC" firstStartedPulling="2025-10-07 21:06:22.550473161 +0000 UTC m=+632.589054321" lastFinishedPulling="2025-10-07 21:06:37.258836086 +0000 UTC m=+647.297417236" observedRunningTime="2025-10-07 21:06:37.836123977 +0000 UTC m=+647.874705127" watchObservedRunningTime="2025-10-07 21:06:37.837439294 +0000 UTC m=+647.876020444" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.862943 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98" podStartSLOduration=2.14837127 podStartE2EDuration="16.862927183s" podCreationTimestamp="2025-10-07 21:06:21 +0000 UTC" firstStartedPulling="2025-10-07 21:06:22.635368878 +0000 UTC m=+632.673950028" lastFinishedPulling="2025-10-07 21:06:37.349924801 +0000 UTC m=+647.388505941" observedRunningTime="2025-10-07 21:06:37.861271546 +0000 UTC m=+647.899852696" watchObservedRunningTime="2025-10-07 21:06:37.862927183 +0000 UTC m=+647.901508323" Oct 07 21:06:37 crc kubenswrapper[4926]: I1007 21:06:37.891362 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-9pr8c" podStartSLOduration=2.13327668 podStartE2EDuration="16.891331436s" podCreationTimestamp="2025-10-07 21:06:21 +0000 UTC" firstStartedPulling="2025-10-07 21:06:22.499471343 +0000 UTC m=+632.538052523" lastFinishedPulling="2025-10-07 21:06:37.257526129 +0000 UTC m=+647.296107279" observedRunningTime="2025-10-07 21:06:37.889082571 +0000 UTC m=+647.927663721" watchObservedRunningTime="2025-10-07 21:06:37.891331436 +0000 UTC m=+647.929912586" Oct 07 21:06:42 crc kubenswrapper[4926]: I1007 21:06:42.268748 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-9vvdj" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.416467 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9"] Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.417884 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.420540 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.435873 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9"] Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.587274 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.587439 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.587578 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgp94\" (UniqueName: \"kubernetes.io/projected/288387f2-b7aa-41d7-a10c-e11e65e054f1-kube-api-access-mgp94\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.688495 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.688611 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.688657 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgp94\" (UniqueName: \"kubernetes.io/projected/288387f2-b7aa-41d7-a10c-e11e65e054f1-kube-api-access-mgp94\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.688962 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.689119 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.706338 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgp94\" (UniqueName: \"kubernetes.io/projected/288387f2-b7aa-41d7-a10c-e11e65e054f1-kube-api-access-mgp94\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:01 crc kubenswrapper[4926]: I1007 21:07:01.736023 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:02 crc kubenswrapper[4926]: I1007 21:07:02.182261 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9"] Oct 07 21:07:02 crc kubenswrapper[4926]: W1007 21:07:02.185562 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod288387f2_b7aa_41d7_a10c_e11e65e054f1.slice/crio-d17dd82419ade5530bd66dda850cd47173576a5e6351d555111ae5cbe7ae90b4 WatchSource:0}: Error finding container d17dd82419ade5530bd66dda850cd47173576a5e6351d555111ae5cbe7ae90b4: Status 404 returned error can't find the container with id d17dd82419ade5530bd66dda850cd47173576a5e6351d555111ae5cbe7ae90b4 Oct 07 21:07:02 crc kubenswrapper[4926]: I1007 21:07:02.905522 4926 generic.go:334] "Generic (PLEG): container finished" podID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerID="599d336f9eff9307c3210de325e73b5838586a0336cd57f0504f6d450eec1b0c" exitCode=0 Oct 07 21:07:02 crc kubenswrapper[4926]: I1007 21:07:02.905617 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" event={"ID":"288387f2-b7aa-41d7-a10c-e11e65e054f1","Type":"ContainerDied","Data":"599d336f9eff9307c3210de325e73b5838586a0336cd57f0504f6d450eec1b0c"} Oct 07 21:07:02 crc kubenswrapper[4926]: I1007 21:07:02.905658 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" event={"ID":"288387f2-b7aa-41d7-a10c-e11e65e054f1","Type":"ContainerStarted","Data":"d17dd82419ade5530bd66dda850cd47173576a5e6351d555111ae5cbe7ae90b4"} Oct 07 21:07:15 crc kubenswrapper[4926]: I1007 21:07:15.996611 4926 generic.go:334] "Generic (PLEG): container finished" podID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerID="96722dd927358d9bf8bd2d476fd6f0c5d2292d60615ea0b531bb6f43c242f5cc" exitCode=0 Oct 07 21:07:15 crc kubenswrapper[4926]: I1007 21:07:15.996694 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" event={"ID":"288387f2-b7aa-41d7-a10c-e11e65e054f1","Type":"ContainerDied","Data":"96722dd927358d9bf8bd2d476fd6f0c5d2292d60615ea0b531bb6f43c242f5cc"} Oct 07 21:07:17 crc kubenswrapper[4926]: I1007 21:07:17.010431 4926 generic.go:334] "Generic (PLEG): container finished" podID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerID="2fe96837ae3ec64d0bff3feb0d372a526b10a6578f0417ff2ca24c9b7de8a1c5" exitCode=0 Oct 07 21:07:17 crc kubenswrapper[4926]: I1007 21:07:17.010613 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" event={"ID":"288387f2-b7aa-41d7-a10c-e11e65e054f1","Type":"ContainerDied","Data":"2fe96837ae3ec64d0bff3feb0d372a526b10a6578f0417ff2ca24c9b7de8a1c5"} Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.327533 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.339763 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-bundle\") pod \"288387f2-b7aa-41d7-a10c-e11e65e054f1\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.339880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgp94\" (UniqueName: \"kubernetes.io/projected/288387f2-b7aa-41d7-a10c-e11e65e054f1-kube-api-access-mgp94\") pod \"288387f2-b7aa-41d7-a10c-e11e65e054f1\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.339911 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-util\") pod \"288387f2-b7aa-41d7-a10c-e11e65e054f1\" (UID: \"288387f2-b7aa-41d7-a10c-e11e65e054f1\") " Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.341043 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-bundle" (OuterVolumeSpecName: "bundle") pod "288387f2-b7aa-41d7-a10c-e11e65e054f1" (UID: "288387f2-b7aa-41d7-a10c-e11e65e054f1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.351536 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/288387f2-b7aa-41d7-a10c-e11e65e054f1-kube-api-access-mgp94" (OuterVolumeSpecName: "kube-api-access-mgp94") pod "288387f2-b7aa-41d7-a10c-e11e65e054f1" (UID: "288387f2-b7aa-41d7-a10c-e11e65e054f1"). InnerVolumeSpecName "kube-api-access-mgp94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.358799 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-util" (OuterVolumeSpecName: "util") pod "288387f2-b7aa-41d7-a10c-e11e65e054f1" (UID: "288387f2-b7aa-41d7-a10c-e11e65e054f1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.441713 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.441758 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgp94\" (UniqueName: \"kubernetes.io/projected/288387f2-b7aa-41d7-a10c-e11e65e054f1-kube-api-access-mgp94\") on node \"crc\" DevicePath \"\"" Oct 07 21:07:18 crc kubenswrapper[4926]: I1007 21:07:18.441777 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/288387f2-b7aa-41d7-a10c-e11e65e054f1-util\") on node \"crc\" DevicePath \"\"" Oct 07 21:07:19 crc kubenswrapper[4926]: I1007 21:07:19.025872 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" event={"ID":"288387f2-b7aa-41d7-a10c-e11e65e054f1","Type":"ContainerDied","Data":"d17dd82419ade5530bd66dda850cd47173576a5e6351d555111ae5cbe7ae90b4"} Oct 07 21:07:19 crc kubenswrapper[4926]: I1007 21:07:19.025924 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d17dd82419ade5530bd66dda850cd47173576a5e6351d555111ae5cbe7ae90b4" Oct 07 21:07:19 crc kubenswrapper[4926]: I1007 21:07:19.025953 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.090741 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn"] Oct 07 21:07:23 crc kubenswrapper[4926]: E1007 21:07:23.091341 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="util" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.091361 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="util" Oct 07 21:07:23 crc kubenswrapper[4926]: E1007 21:07:23.091384 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="extract" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.091397 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="extract" Oct 07 21:07:23 crc kubenswrapper[4926]: E1007 21:07:23.091426 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="pull" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.091441 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="pull" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.091596 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="288387f2-b7aa-41d7-a10c-e11e65e054f1" containerName="extract" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.092166 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.094544 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.095532 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.097091 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xjjfk" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.107944 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn"] Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.112430 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf98m\" (UniqueName: \"kubernetes.io/projected/c6ce449c-47a2-4bf7-9556-c3b72d502494-kube-api-access-jf98m\") pod \"nmstate-operator-858ddd8f98-vqnbn\" (UID: \"c6ce449c-47a2-4bf7-9556-c3b72d502494\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.213996 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf98m\" (UniqueName: \"kubernetes.io/projected/c6ce449c-47a2-4bf7-9556-c3b72d502494-kube-api-access-jf98m\") pod \"nmstate-operator-858ddd8f98-vqnbn\" (UID: \"c6ce449c-47a2-4bf7-9556-c3b72d502494\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.240435 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf98m\" (UniqueName: \"kubernetes.io/projected/c6ce449c-47a2-4bf7-9556-c3b72d502494-kube-api-access-jf98m\") pod \"nmstate-operator-858ddd8f98-vqnbn\" (UID: \"c6ce449c-47a2-4bf7-9556-c3b72d502494\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.420886 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" Oct 07 21:07:23 crc kubenswrapper[4926]: I1007 21:07:23.677490 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn"] Oct 07 21:07:24 crc kubenswrapper[4926]: I1007 21:07:24.057536 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" event={"ID":"c6ce449c-47a2-4bf7-9556-c3b72d502494","Type":"ContainerStarted","Data":"54dcc4c0c33b0149d44909bbf14d3a664bd6c05c18f0c1f8a6b4b96dd31466dc"} Oct 07 21:07:26 crc kubenswrapper[4926]: I1007 21:07:26.070709 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" event={"ID":"c6ce449c-47a2-4bf7-9556-c3b72d502494","Type":"ContainerStarted","Data":"1372140f9f5fb49164d88f29da981c54df7fe7fd7fbc37d85ad180f067347f7e"} Oct 07 21:07:32 crc kubenswrapper[4926]: I1007 21:07:32.980604 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vqnbn" podStartSLOduration=7.985622849 podStartE2EDuration="9.980581986s" podCreationTimestamp="2025-10-07 21:07:23 +0000 UTC" firstStartedPulling="2025-10-07 21:07:23.695725638 +0000 UTC m=+693.734306788" lastFinishedPulling="2025-10-07 21:07:25.690684775 +0000 UTC m=+695.729265925" observedRunningTime="2025-10-07 21:07:26.087653168 +0000 UTC m=+696.126234368" watchObservedRunningTime="2025-10-07 21:07:32.980581986 +0000 UTC m=+703.019163146" Oct 07 21:07:32 crc kubenswrapper[4926]: I1007 21:07:32.981738 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9"] Oct 07 21:07:32 crc kubenswrapper[4926]: I1007 21:07:32.983056 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" Oct 07 21:07:32 crc kubenswrapper[4926]: I1007 21:07:32.984860 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nxv48" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.003059 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.045390 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.046265 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:33 crc kubenswrapper[4926]: W1007 21:07:33.048269 4926 reflector.go:561] object-"openshift-nmstate"/"openshift-nmstate-webhook": failed to list *v1.Secret: secrets "openshift-nmstate-webhook" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Oct 07 21:07:33 crc kubenswrapper[4926]: E1007 21:07:33.048334 4926 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"openshift-nmstate-webhook\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-nmstate-webhook\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.067065 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.070152 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-lswbm"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.070868 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.159088 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.160163 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.162458 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.162463 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167230 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5s4p\" (UniqueName: \"kubernetes.io/projected/aca14e9c-b601-4092-b7a5-72a80506b27e-kube-api-access-g5s4p\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167456 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aca14e9c-b601-4092-b7a5-72a80506b27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167592 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86ld9\" (UniqueName: \"kubernetes.io/projected/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-kube-api-access-86ld9\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167700 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-nmstate-lock\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167796 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swv42\" (UniqueName: \"kubernetes.io/projected/8daf114a-72f8-44e9-ac14-166bdf8e73cb-kube-api-access-swv42\") pod \"nmstate-metrics-fdff9cb8d-mczl9\" (UID: \"8daf114a-72f8-44e9-ac14-166bdf8e73cb\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167891 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-ovs-socket\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.167991 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-dbus-socket\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.168225 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2gt4b" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.170377 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.209430 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.209488 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269642 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsbtg\" (UniqueName: \"kubernetes.io/projected/7b429d99-24bf-4fee-81fb-eb1bc76d9788-kube-api-access-xsbtg\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269717 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b429d99-24bf-4fee-81fb-eb1bc76d9788-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269755 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5s4p\" (UniqueName: \"kubernetes.io/projected/aca14e9c-b601-4092-b7a5-72a80506b27e-kube-api-access-g5s4p\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269798 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aca14e9c-b601-4092-b7a5-72a80506b27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269842 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-nmstate-lock\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269864 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86ld9\" (UniqueName: \"kubernetes.io/projected/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-kube-api-access-86ld9\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269888 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swv42\" (UniqueName: \"kubernetes.io/projected/8daf114a-72f8-44e9-ac14-166bdf8e73cb-kube-api-access-swv42\") pod \"nmstate-metrics-fdff9cb8d-mczl9\" (UID: \"8daf114a-72f8-44e9-ac14-166bdf8e73cb\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-ovs-socket\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269932 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-nmstate-lock\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.269945 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-dbus-socket\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.270031 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7b429d99-24bf-4fee-81fb-eb1bc76d9788-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.270128 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-ovs-socket\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.270185 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-dbus-socket\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.292886 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86ld9\" (UniqueName: \"kubernetes.io/projected/8c97c92f-f0b8-48ec-b5f7-6594bd06c948-kube-api-access-86ld9\") pod \"nmstate-handler-lswbm\" (UID: \"8c97c92f-f0b8-48ec-b5f7-6594bd06c948\") " pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.293133 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swv42\" (UniqueName: \"kubernetes.io/projected/8daf114a-72f8-44e9-ac14-166bdf8e73cb-kube-api-access-swv42\") pod \"nmstate-metrics-fdff9cb8d-mczl9\" (UID: \"8daf114a-72f8-44e9-ac14-166bdf8e73cb\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.310351 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5s4p\" (UniqueName: \"kubernetes.io/projected/aca14e9c-b601-4092-b7a5-72a80506b27e-kube-api-access-g5s4p\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.338618 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.370838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b429d99-24bf-4fee-81fb-eb1bc76d9788-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.370964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7b429d99-24bf-4fee-81fb-eb1bc76d9788-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.371005 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsbtg\" (UniqueName: \"kubernetes.io/projected/7b429d99-24bf-4fee-81fb-eb1bc76d9788-kube-api-access-xsbtg\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.371144 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-78dc758475-557ll"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.372125 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.372428 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7b429d99-24bf-4fee-81fb-eb1bc76d9788-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.376291 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b429d99-24bf-4fee-81fb-eb1bc76d9788-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.385041 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78dc758475-557ll"] Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.387178 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.402600 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsbtg\" (UniqueName: \"kubernetes.io/projected/7b429d99-24bf-4fee-81fb-eb1bc76d9788-kube-api-access-xsbtg\") pod \"nmstate-console-plugin-6b874cbd85-wdx8h\" (UID: \"7b429d99-24bf-4fee-81fb-eb1bc76d9788\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.472601 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-service-ca\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.472900 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-serving-cert\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.473000 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxltn\" (UniqueName: \"kubernetes.io/projected/f4666c1d-74ee-46e2-a76e-755ea5945c59-kube-api-access-gxltn\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.473029 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-trusted-ca-bundle\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.473061 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-config\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.473081 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-oauth-config\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.473117 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-oauth-serving-cert\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.473356 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574488 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-config\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574527 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-oauth-config\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574576 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-oauth-serving-cert\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574622 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-service-ca\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574640 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-serving-cert\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574676 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxltn\" (UniqueName: \"kubernetes.io/projected/f4666c1d-74ee-46e2-a76e-755ea5945c59-kube-api-access-gxltn\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.574693 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-trusted-ca-bundle\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.575611 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-oauth-serving-cert\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.576188 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-trusted-ca-bundle\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.576256 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-service-ca\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.578010 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-config\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.579692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-oauth-config\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.579788 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4666c1d-74ee-46e2-a76e-755ea5945c59-console-serving-cert\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.592075 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxltn\" (UniqueName: \"kubernetes.io/projected/f4666c1d-74ee-46e2-a76e-755ea5945c59-kube-api-access-gxltn\") pod \"console-78dc758475-557ll\" (UID: \"f4666c1d-74ee-46e2-a76e-755ea5945c59\") " pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.657158 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h"] Oct 07 21:07:33 crc kubenswrapper[4926]: W1007 21:07:33.669419 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b429d99_24bf_4fee_81fb_eb1bc76d9788.slice/crio-110ca11a48c163d1ca76977429682f9a405f730b532e849db90923cc0c7c9f78 WatchSource:0}: Error finding container 110ca11a48c163d1ca76977429682f9a405f730b532e849db90923cc0c7c9f78: Status 404 returned error can't find the container with id 110ca11a48c163d1ca76977429682f9a405f730b532e849db90923cc0c7c9f78 Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.735561 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:33 crc kubenswrapper[4926]: I1007 21:07:33.751522 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9"] Oct 07 21:07:33 crc kubenswrapper[4926]: W1007 21:07:33.759051 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8daf114a_72f8_44e9_ac14_166bdf8e73cb.slice/crio-9ee9898d398e47b5ee1be1c5205252fbb2f23256dc4d1a3c729d3b5c1e87f348 WatchSource:0}: Error finding container 9ee9898d398e47b5ee1be1c5205252fbb2f23256dc4d1a3c729d3b5c1e87f348: Status 404 returned error can't find the container with id 9ee9898d398e47b5ee1be1c5205252fbb2f23256dc4d1a3c729d3b5c1e87f348 Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.125591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" event={"ID":"7b429d99-24bf-4fee-81fb-eb1bc76d9788","Type":"ContainerStarted","Data":"110ca11a48c163d1ca76977429682f9a405f730b532e849db90923cc0c7c9f78"} Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.126797 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" event={"ID":"8daf114a-72f8-44e9-ac14-166bdf8e73cb","Type":"ContainerStarted","Data":"9ee9898d398e47b5ee1be1c5205252fbb2f23256dc4d1a3c729d3b5c1e87f348"} Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.127591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lswbm" event={"ID":"8c97c92f-f0b8-48ec-b5f7-6594bd06c948","Type":"ContainerStarted","Data":"fec351db0defe3ff63fd663964e03332b193481e0502c2c39c590ccb74276878"} Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.170045 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78dc758475-557ll"] Oct 07 21:07:34 crc kubenswrapper[4926]: W1007 21:07:34.178183 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4666c1d_74ee_46e2_a76e_755ea5945c59.slice/crio-7686fd3e17eb9aa62a3d4d7ddd8908dee01d84d03571f663ae117222a9c3bc30 WatchSource:0}: Error finding container 7686fd3e17eb9aa62a3d4d7ddd8908dee01d84d03571f663ae117222a9c3bc30: Status 404 returned error can't find the container with id 7686fd3e17eb9aa62a3d4d7ddd8908dee01d84d03571f663ae117222a9c3bc30 Oct 07 21:07:34 crc kubenswrapper[4926]: E1007 21:07:34.270572 4926 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: failed to sync secret cache: timed out waiting for the condition Oct 07 21:07:34 crc kubenswrapper[4926]: E1007 21:07:34.270755 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aca14e9c-b601-4092-b7a5-72a80506b27e-tls-key-pair podName:aca14e9c-b601-4092-b7a5-72a80506b27e nodeName:}" failed. No retries permitted until 2025-10-07 21:07:34.770733683 +0000 UTC m=+704.809314833 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/aca14e9c-b601-4092-b7a5-72a80506b27e-tls-key-pair") pod "nmstate-webhook-6cdbc54649-r4nvj" (UID: "aca14e9c-b601-4092-b7a5-72a80506b27e") : failed to sync secret cache: timed out waiting for the condition Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.559231 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.788361 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aca14e9c-b601-4092-b7a5-72a80506b27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.793774 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aca14e9c-b601-4092-b7a5-72a80506b27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r4nvj\" (UID: \"aca14e9c-b601-4092-b7a5-72a80506b27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:34 crc kubenswrapper[4926]: I1007 21:07:34.863598 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:35 crc kubenswrapper[4926]: I1007 21:07:35.141083 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78dc758475-557ll" event={"ID":"f4666c1d-74ee-46e2-a76e-755ea5945c59","Type":"ContainerStarted","Data":"1a8c60925064352760a31e16587fd4c683c6e95db9259446ddd71be4a11be0a0"} Oct 07 21:07:35 crc kubenswrapper[4926]: I1007 21:07:35.141651 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78dc758475-557ll" event={"ID":"f4666c1d-74ee-46e2-a76e-755ea5945c59","Type":"ContainerStarted","Data":"7686fd3e17eb9aa62a3d4d7ddd8908dee01d84d03571f663ae117222a9c3bc30"} Oct 07 21:07:35 crc kubenswrapper[4926]: I1007 21:07:35.328334 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-78dc758475-557ll" podStartSLOduration=2.3283124 podStartE2EDuration="2.3283124s" podCreationTimestamp="2025-10-07 21:07:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:07:35.167325846 +0000 UTC m=+705.205906996" watchObservedRunningTime="2025-10-07 21:07:35.3283124 +0000 UTC m=+705.366893550" Oct 07 21:07:35 crc kubenswrapper[4926]: I1007 21:07:35.331352 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj"] Oct 07 21:07:36 crc kubenswrapper[4926]: I1007 21:07:36.156934 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" event={"ID":"aca14e9c-b601-4092-b7a5-72a80506b27e","Type":"ContainerStarted","Data":"e38c726c02d31c6f1020d89cb1c5f43614802f342bcc6f539213c570a9e93f2b"} Oct 07 21:07:37 crc kubenswrapper[4926]: I1007 21:07:37.171383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lswbm" event={"ID":"8c97c92f-f0b8-48ec-b5f7-6594bd06c948","Type":"ContainerStarted","Data":"5750f13919c49df32bbd06134c8436deb933f68aae18fcbb493b24e2a94eb639"} Oct 07 21:07:37 crc kubenswrapper[4926]: I1007 21:07:37.172262 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:37 crc kubenswrapper[4926]: I1007 21:07:37.176170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" event={"ID":"7b429d99-24bf-4fee-81fb-eb1bc76d9788","Type":"ContainerStarted","Data":"686d96ed1d6376241a8d159d72ff840ead6b1f96765146a6533ccb43caef7a5a"} Oct 07 21:07:37 crc kubenswrapper[4926]: I1007 21:07:37.178549 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" event={"ID":"8daf114a-72f8-44e9-ac14-166bdf8e73cb","Type":"ContainerStarted","Data":"89d5f648107fe3053f0c118d6b33d63a88a86c3f98df0b0c2749e11525e7b91d"} Oct 07 21:07:37 crc kubenswrapper[4926]: I1007 21:07:37.195607 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-lswbm" podStartSLOduration=2.084588659 podStartE2EDuration="5.195592034s" podCreationTimestamp="2025-10-07 21:07:32 +0000 UTC" firstStartedPulling="2025-10-07 21:07:33.41291394 +0000 UTC m=+703.451495090" lastFinishedPulling="2025-10-07 21:07:36.523917315 +0000 UTC m=+706.562498465" observedRunningTime="2025-10-07 21:07:37.191521198 +0000 UTC m=+707.230102368" watchObservedRunningTime="2025-10-07 21:07:37.195592034 +0000 UTC m=+707.234173184" Oct 07 21:07:37 crc kubenswrapper[4926]: I1007 21:07:37.217008 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wdx8h" podStartSLOduration=1.374986665 podStartE2EDuration="4.216987906s" podCreationTimestamp="2025-10-07 21:07:33 +0000 UTC" firstStartedPulling="2025-10-07 21:07:33.676623342 +0000 UTC m=+703.715204492" lastFinishedPulling="2025-10-07 21:07:36.518624533 +0000 UTC m=+706.557205733" observedRunningTime="2025-10-07 21:07:37.216660247 +0000 UTC m=+707.255241437" watchObservedRunningTime="2025-10-07 21:07:37.216987906 +0000 UTC m=+707.255569066" Oct 07 21:07:38 crc kubenswrapper[4926]: I1007 21:07:38.195066 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" event={"ID":"aca14e9c-b601-4092-b7a5-72a80506b27e","Type":"ContainerStarted","Data":"d5efecd1b27ba50a641ba1c5d16ef927582a38f10bd981f673de6078205ee894"} Oct 07 21:07:38 crc kubenswrapper[4926]: I1007 21:07:38.195889 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:07:38 crc kubenswrapper[4926]: I1007 21:07:38.222827 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" podStartSLOduration=5.148129758 podStartE2EDuration="6.222806843s" podCreationTimestamp="2025-10-07 21:07:32 +0000 UTC" firstStartedPulling="2025-10-07 21:07:36.067253114 +0000 UTC m=+706.105834274" lastFinishedPulling="2025-10-07 21:07:37.141930179 +0000 UTC m=+707.180511359" observedRunningTime="2025-10-07 21:07:38.220474276 +0000 UTC m=+708.259055496" watchObservedRunningTime="2025-10-07 21:07:38.222806843 +0000 UTC m=+708.261388003" Oct 07 21:07:40 crc kubenswrapper[4926]: I1007 21:07:40.213951 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" event={"ID":"8daf114a-72f8-44e9-ac14-166bdf8e73cb","Type":"ContainerStarted","Data":"6ed13725e2eb265b532f90d2a16b4abbe69ef8506dc3fdb20c670261cee54bf4"} Oct 07 21:07:43 crc kubenswrapper[4926]: I1007 21:07:43.426360 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-lswbm" Oct 07 21:07:43 crc kubenswrapper[4926]: I1007 21:07:43.457237 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-mczl9" podStartSLOduration=5.958347258 podStartE2EDuration="11.457173025s" podCreationTimestamp="2025-10-07 21:07:32 +0000 UTC" firstStartedPulling="2025-10-07 21:07:33.762417545 +0000 UTC m=+703.800998715" lastFinishedPulling="2025-10-07 21:07:39.261243322 +0000 UTC m=+709.299824482" observedRunningTime="2025-10-07 21:07:40.245870333 +0000 UTC m=+710.284451533" watchObservedRunningTime="2025-10-07 21:07:43.457173025 +0000 UTC m=+713.495754245" Oct 07 21:07:43 crc kubenswrapper[4926]: I1007 21:07:43.736974 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:43 crc kubenswrapper[4926]: I1007 21:07:43.737399 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:43 crc kubenswrapper[4926]: I1007 21:07:43.745350 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:44 crc kubenswrapper[4926]: I1007 21:07:44.254554 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-78dc758475-557ll" Oct 07 21:07:44 crc kubenswrapper[4926]: I1007 21:07:44.355173 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ggd2q"] Oct 07 21:07:54 crc kubenswrapper[4926]: I1007 21:07:54.872615 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r4nvj" Oct 07 21:08:03 crc kubenswrapper[4926]: I1007 21:08:03.209822 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:08:03 crc kubenswrapper[4926]: I1007 21:08:03.210427 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.418144 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ggd2q" podUID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" containerName="console" containerID="cri-o://de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291" gracePeriod=15 Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.853076 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ggd2q_0a6d69ac-d24d-4e99-9ea8-ffafe15885ec/console/0.log" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.853653 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947290 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-config\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947336 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-service-ca\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947385 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-oauth-config\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947431 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfvlm\" (UniqueName: \"kubernetes.io/projected/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-kube-api-access-mfvlm\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947465 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-oauth-serving-cert\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947495 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-trusted-ca-bundle\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.947525 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-serving-cert\") pod \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\" (UID: \"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec\") " Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.948269 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-service-ca" (OuterVolumeSpecName: "service-ca") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.948341 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-config" (OuterVolumeSpecName: "console-config") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.948719 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.948774 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.966505 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.967493 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:08:09 crc kubenswrapper[4926]: I1007 21:08:09.969168 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-kube-api-access-mfvlm" (OuterVolumeSpecName: "kube-api-access-mfvlm") pod "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" (UID: "0a6d69ac-d24d-4e99-9ea8-ffafe15885ec"). InnerVolumeSpecName "kube-api-access-mfvlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049071 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049168 4926 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049185 4926 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049214 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfvlm\" (UniqueName: \"kubernetes.io/projected/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-kube-api-access-mfvlm\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049225 4926 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049235 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.049346 4926 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.436917 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ggd2q_0a6d69ac-d24d-4e99-9ea8-ffafe15885ec/console/0.log" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.436955 4926 generic.go:334] "Generic (PLEG): container finished" podID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" containerID="de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291" exitCode=2 Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.436978 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ggd2q" event={"ID":"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec","Type":"ContainerDied","Data":"de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291"} Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.437001 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ggd2q" event={"ID":"0a6d69ac-d24d-4e99-9ea8-ffafe15885ec","Type":"ContainerDied","Data":"c98d2abad18ece47103576e3aa9d6828cf47b3f6e4ce1837ff9555f3f0c30a1c"} Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.437016 4926 scope.go:117] "RemoveContainer" containerID="de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.437112 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ggd2q" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.461637 4926 scope.go:117] "RemoveContainer" containerID="de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291" Oct 07 21:08:10 crc kubenswrapper[4926]: E1007 21:08:10.462001 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291\": container with ID starting with de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291 not found: ID does not exist" containerID="de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.462026 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291"} err="failed to get container status \"de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291\": rpc error: code = NotFound desc = could not find container \"de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291\": container with ID starting with de1a3997b998ce68c3ff1ce9ef066ab457ef424c6303cf62ad1fc48dd1449291 not found: ID does not exist" Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.465314 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ggd2q"] Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.466896 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ggd2q"] Oct 07 21:08:10 crc kubenswrapper[4926]: I1007 21:08:10.700302 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" path="/var/lib/kubelet/pods/0a6d69ac-d24d-4e99-9ea8-ffafe15885ec/volumes" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.722744 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz"] Oct 07 21:08:11 crc kubenswrapper[4926]: E1007 21:08:11.723481 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" containerName="console" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.723503 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" containerName="console" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.723698 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a6d69ac-d24d-4e99-9ea8-ffafe15885ec" containerName="console" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.725563 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.728651 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.741493 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz"] Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.779554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.779669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.779714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wplf2\" (UniqueName: \"kubernetes.io/projected/b5d6db52-5370-492c-8641-7fa8e5601d4a-kube-api-access-wplf2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.881637 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.881763 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.881806 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wplf2\" (UniqueName: \"kubernetes.io/projected/b5d6db52-5370-492c-8641-7fa8e5601d4a-kube-api-access-wplf2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.882525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.882922 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:11 crc kubenswrapper[4926]: I1007 21:08:11.907578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wplf2\" (UniqueName: \"kubernetes.io/projected/b5d6db52-5370-492c-8641-7fa8e5601d4a-kube-api-access-wplf2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:12 crc kubenswrapper[4926]: I1007 21:08:12.062882 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:12 crc kubenswrapper[4926]: I1007 21:08:12.269803 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz"] Oct 07 21:08:12 crc kubenswrapper[4926]: I1007 21:08:12.457491 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" event={"ID":"b5d6db52-5370-492c-8641-7fa8e5601d4a","Type":"ContainerStarted","Data":"350fdee3dcead982f3cc56d22c2b11572b79a4f760db30a590442a30c7a09861"} Oct 07 21:08:12 crc kubenswrapper[4926]: I1007 21:08:12.457543 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" event={"ID":"b5d6db52-5370-492c-8641-7fa8e5601d4a","Type":"ContainerStarted","Data":"b8df271a3d84ee8add091394638581865ac7c3f28702a790b3cb25373fdad3f7"} Oct 07 21:08:13 crc kubenswrapper[4926]: I1007 21:08:13.470586 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerID="350fdee3dcead982f3cc56d22c2b11572b79a4f760db30a590442a30c7a09861" exitCode=0 Oct 07 21:08:13 crc kubenswrapper[4926]: I1007 21:08:13.470802 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" event={"ID":"b5d6db52-5370-492c-8641-7fa8e5601d4a","Type":"ContainerDied","Data":"350fdee3dcead982f3cc56d22c2b11572b79a4f760db30a590442a30c7a09861"} Oct 07 21:08:15 crc kubenswrapper[4926]: I1007 21:08:15.489766 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerID="7fe615f3dcf8499ce6254beb58e0ec78a8b66ee7d2a7264a892d646500ce7b54" exitCode=0 Oct 07 21:08:15 crc kubenswrapper[4926]: I1007 21:08:15.489819 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" event={"ID":"b5d6db52-5370-492c-8641-7fa8e5601d4a","Type":"ContainerDied","Data":"7fe615f3dcf8499ce6254beb58e0ec78a8b66ee7d2a7264a892d646500ce7b54"} Oct 07 21:08:16 crc kubenswrapper[4926]: I1007 21:08:16.500327 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerID="4f11ba332d2b888fc50a57c4be909a381b98d425989e5e300f654ee34d6796d8" exitCode=0 Oct 07 21:08:16 crc kubenswrapper[4926]: I1007 21:08:16.500377 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" event={"ID":"b5d6db52-5370-492c-8641-7fa8e5601d4a","Type":"ContainerDied","Data":"4f11ba332d2b888fc50a57c4be909a381b98d425989e5e300f654ee34d6796d8"} Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.213608 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k87k7"] Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.213908 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" podUID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" containerName="controller-manager" containerID="cri-o://2de8240ecf0a5bb884042a1f01bf268880724f81a19442615eac4f9f323a6c1e" gracePeriod=30 Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.312279 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb"] Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.312516 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" podUID="6d6919d5-336c-452e-94dd-f37223a44f2c" containerName="route-controller-manager" containerID="cri-o://fb95c0f4ed1744704088e003607b31297eb4fa254a5ab45d7893d9b012ea7c8a" gracePeriod=30 Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.506809 4926 generic.go:334] "Generic (PLEG): container finished" podID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" containerID="2de8240ecf0a5bb884042a1f01bf268880724f81a19442615eac4f9f323a6c1e" exitCode=0 Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.506959 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" event={"ID":"b37fc498-33f9-4968-8c6a-1cfb00e123b8","Type":"ContainerDied","Data":"2de8240ecf0a5bb884042a1f01bf268880724f81a19442615eac4f9f323a6c1e"} Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.508558 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d6919d5-336c-452e-94dd-f37223a44f2c" containerID="fb95c0f4ed1744704088e003607b31297eb4fa254a5ab45d7893d9b012ea7c8a" exitCode=0 Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.508667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" event={"ID":"6d6919d5-336c-452e-94dd-f37223a44f2c","Type":"ContainerDied","Data":"fb95c0f4ed1744704088e003607b31297eb4fa254a5ab45d7893d9b012ea7c8a"} Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.669797 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.672941 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.739132 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774734 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-client-ca\") pod \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774801 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkjvr\" (UniqueName: \"kubernetes.io/projected/b37fc498-33f9-4968-8c6a-1cfb00e123b8-kube-api-access-hkjvr\") pod \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774824 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d6919d5-336c-452e-94dd-f37223a44f2c-serving-cert\") pod \"6d6919d5-336c-452e-94dd-f37223a44f2c\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774879 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-config\") pod \"6d6919d5-336c-452e-94dd-f37223a44f2c\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774914 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-client-ca\") pod \"6d6919d5-336c-452e-94dd-f37223a44f2c\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774942 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-proxy-ca-bundles\") pod \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.774956 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37fc498-33f9-4968-8c6a-1cfb00e123b8-serving-cert\") pod \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.776156 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qvcn\" (UniqueName: \"kubernetes.io/projected/6d6919d5-336c-452e-94dd-f37223a44f2c-kube-api-access-6qvcn\") pod \"6d6919d5-336c-452e-94dd-f37223a44f2c\" (UID: \"6d6919d5-336c-452e-94dd-f37223a44f2c\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.776383 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-config\") pod \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\" (UID: \"b37fc498-33f9-4968-8c6a-1cfb00e123b8\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.777242 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-config" (OuterVolumeSpecName: "config") pod "6d6919d5-336c-452e-94dd-f37223a44f2c" (UID: "6d6919d5-336c-452e-94dd-f37223a44f2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.777339 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-config" (OuterVolumeSpecName: "config") pod "b37fc498-33f9-4968-8c6a-1cfb00e123b8" (UID: "b37fc498-33f9-4968-8c6a-1cfb00e123b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.777705 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-client-ca" (OuterVolumeSpecName: "client-ca") pod "6d6919d5-336c-452e-94dd-f37223a44f2c" (UID: "6d6919d5-336c-452e-94dd-f37223a44f2c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.778069 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b37fc498-33f9-4968-8c6a-1cfb00e123b8" (UID: "b37fc498-33f9-4968-8c6a-1cfb00e123b8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.778087 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-client-ca" (OuterVolumeSpecName: "client-ca") pod "b37fc498-33f9-4968-8c6a-1cfb00e123b8" (UID: "b37fc498-33f9-4968-8c6a-1cfb00e123b8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.790421 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d6919d5-336c-452e-94dd-f37223a44f2c-kube-api-access-6qvcn" (OuterVolumeSpecName: "kube-api-access-6qvcn") pod "6d6919d5-336c-452e-94dd-f37223a44f2c" (UID: "6d6919d5-336c-452e-94dd-f37223a44f2c"). InnerVolumeSpecName "kube-api-access-6qvcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.790723 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d6919d5-336c-452e-94dd-f37223a44f2c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6d6919d5-336c-452e-94dd-f37223a44f2c" (UID: "6d6919d5-336c-452e-94dd-f37223a44f2c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.791111 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b37fc498-33f9-4968-8c6a-1cfb00e123b8-kube-api-access-hkjvr" (OuterVolumeSpecName: "kube-api-access-hkjvr") pod "b37fc498-33f9-4968-8c6a-1cfb00e123b8" (UID: "b37fc498-33f9-4968-8c6a-1cfb00e123b8"). InnerVolumeSpecName "kube-api-access-hkjvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.791585 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b37fc498-33f9-4968-8c6a-1cfb00e123b8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b37fc498-33f9-4968-8c6a-1cfb00e123b8" (UID: "b37fc498-33f9-4968-8c6a-1cfb00e123b8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877301 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-util\") pod \"b5d6db52-5370-492c-8641-7fa8e5601d4a\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877389 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-bundle\") pod \"b5d6db52-5370-492c-8641-7fa8e5601d4a\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wplf2\" (UniqueName: \"kubernetes.io/projected/b5d6db52-5370-492c-8641-7fa8e5601d4a-kube-api-access-wplf2\") pod \"b5d6db52-5370-492c-8641-7fa8e5601d4a\" (UID: \"b5d6db52-5370-492c-8641-7fa8e5601d4a\") " Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877875 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877903 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d6919d5-336c-452e-94dd-f37223a44f2c-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877922 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877941 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37fc498-33f9-4968-8c6a-1cfb00e123b8-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877958 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qvcn\" (UniqueName: \"kubernetes.io/projected/6d6919d5-336c-452e-94dd-f37223a44f2c-kube-api-access-6qvcn\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877975 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.877991 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37fc498-33f9-4968-8c6a-1cfb00e123b8-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.878007 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkjvr\" (UniqueName: \"kubernetes.io/projected/b37fc498-33f9-4968-8c6a-1cfb00e123b8-kube-api-access-hkjvr\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.878055 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d6919d5-336c-452e-94dd-f37223a44f2c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.879907 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-bundle" (OuterVolumeSpecName: "bundle") pod "b5d6db52-5370-492c-8641-7fa8e5601d4a" (UID: "b5d6db52-5370-492c-8641-7fa8e5601d4a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.881900 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d6db52-5370-492c-8641-7fa8e5601d4a-kube-api-access-wplf2" (OuterVolumeSpecName: "kube-api-access-wplf2") pod "b5d6db52-5370-492c-8641-7fa8e5601d4a" (UID: "b5d6db52-5370-492c-8641-7fa8e5601d4a"). InnerVolumeSpecName "kube-api-access-wplf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.890045 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-util" (OuterVolumeSpecName: "util") pod "b5d6db52-5370-492c-8641-7fa8e5601d4a" (UID: "b5d6db52-5370-492c-8641-7fa8e5601d4a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.978986 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-util\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.979017 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5d6db52-5370-492c-8641-7fa8e5601d4a-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:17 crc kubenswrapper[4926]: I1007 21:08:17.979026 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wplf2\" (UniqueName: \"kubernetes.io/projected/b5d6db52-5370-492c-8641-7fa8e5601d4a-kube-api-access-wplf2\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.515838 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" event={"ID":"b5d6db52-5370-492c-8641-7fa8e5601d4a","Type":"ContainerDied","Data":"b8df271a3d84ee8add091394638581865ac7c3f28702a790b3cb25373fdad3f7"} Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.516325 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8df271a3d84ee8add091394638581865ac7c3f28702a790b3cb25373fdad3f7" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.515883 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.517512 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.517521 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k87k7" event={"ID":"b37fc498-33f9-4968-8c6a-1cfb00e123b8","Type":"ContainerDied","Data":"c93912449079edb50720d1b52345525193e72ad3920202281eeee277e6398686"} Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.517569 4926 scope.go:117] "RemoveContainer" containerID="2de8240ecf0a5bb884042a1f01bf268880724f81a19442615eac4f9f323a6c1e" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.529372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" event={"ID":"6d6919d5-336c-452e-94dd-f37223a44f2c","Type":"ContainerDied","Data":"7cbb8cd844bcd525427c03df9581163f590706f2b2e1e8aceedf04d97d3b354d"} Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.529452 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.551285 4926 scope.go:117] "RemoveContainer" containerID="fb95c0f4ed1744704088e003607b31297eb4fa254a5ab45d7893d9b012ea7c8a" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.556016 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k87k7"] Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.563724 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k87k7"] Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.573543 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb"] Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.576480 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-85zcb"] Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.692055 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d6919d5-336c-452e-94dd-f37223a44f2c" path="/var/lib/kubelet/pods/6d6919d5-336c-452e-94dd-f37223a44f2c/volumes" Oct 07 21:08:18 crc kubenswrapper[4926]: I1007 21:08:18.692747 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" path="/var/lib/kubelet/pods/b37fc498-33f9-4968-8c6a-1cfb00e123b8/volumes" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.047779 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj"] Oct 07 21:08:19 crc kubenswrapper[4926]: E1007 21:08:19.048154 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6919d5-336c-452e-94dd-f37223a44f2c" containerName="route-controller-manager" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048174 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6919d5-336c-452e-94dd-f37223a44f2c" containerName="route-controller-manager" Oct 07 21:08:19 crc kubenswrapper[4926]: E1007 21:08:19.048186 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="util" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048206 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="util" Oct 07 21:08:19 crc kubenswrapper[4926]: E1007 21:08:19.048224 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="extract" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048231 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="extract" Oct 07 21:08:19 crc kubenswrapper[4926]: E1007 21:08:19.048244 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" containerName="controller-manager" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048252 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" containerName="controller-manager" Oct 07 21:08:19 crc kubenswrapper[4926]: E1007 21:08:19.048264 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="pull" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048270 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="pull" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048385 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6919d5-336c-452e-94dd-f37223a44f2c" containerName="route-controller-manager" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048400 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d6db52-5370-492c-8641-7fa8e5601d4a" containerName="extract" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.048412 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b37fc498-33f9-4968-8c6a-1cfb00e123b8" containerName="controller-manager" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.049028 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.052014 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.052539 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.052839 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.053060 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.053489 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.058543 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.062285 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d5766f97d-tbnns"] Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.063522 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.066292 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d5766f97d-tbnns"] Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.068216 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.069116 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.069890 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.070079 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.070314 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.070509 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.070732 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj"] Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.080567 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.099756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbm75\" (UniqueName: \"kubernetes.io/projected/7d4d7e08-b700-461a-9962-1119d0902eb6-kube-api-access-zbm75\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.099834 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d4d7e08-b700-461a-9962-1119d0902eb6-serving-cert\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.099870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4d7e08-b700-461a-9962-1119d0902eb6-config\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.099948 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d4d7e08-b700-461a-9962-1119d0902eb6-client-ca\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201564 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-client-ca\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201639 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb7lx\" (UniqueName: \"kubernetes.io/projected/f7a923dd-1bed-4748-8774-c0efa6640e52-kube-api-access-pb7lx\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201688 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-config\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201724 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d4d7e08-b700-461a-9962-1119d0902eb6-client-ca\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201751 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-proxy-ca-bundles\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201785 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbm75\" (UniqueName: \"kubernetes.io/projected/7d4d7e08-b700-461a-9962-1119d0902eb6-kube-api-access-zbm75\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201824 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d4d7e08-b700-461a-9962-1119d0902eb6-serving-cert\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201849 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4d7e08-b700-461a-9962-1119d0902eb6-config\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.201870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a923dd-1bed-4748-8774-c0efa6640e52-serving-cert\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.205367 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d4d7e08-b700-461a-9962-1119d0902eb6-client-ca\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.205420 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4d7e08-b700-461a-9962-1119d0902eb6-config\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.207669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d4d7e08-b700-461a-9962-1119d0902eb6-serving-cert\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.218772 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbm75\" (UniqueName: \"kubernetes.io/projected/7d4d7e08-b700-461a-9962-1119d0902eb6-kube-api-access-zbm75\") pod \"route-controller-manager-86cbbcd584-2djtj\" (UID: \"7d4d7e08-b700-461a-9962-1119d0902eb6\") " pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.303250 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-config\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.303336 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-proxy-ca-bundles\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.303412 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a923dd-1bed-4748-8774-c0efa6640e52-serving-cert\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.303500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-client-ca\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.303553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb7lx\" (UniqueName: \"kubernetes.io/projected/f7a923dd-1bed-4748-8774-c0efa6640e52-kube-api-access-pb7lx\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.305533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-client-ca\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.305725 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-proxy-ca-bundles\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.305900 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a923dd-1bed-4748-8774-c0efa6640e52-config\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.309300 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a923dd-1bed-4748-8774-c0efa6640e52-serving-cert\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.327942 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb7lx\" (UniqueName: \"kubernetes.io/projected/f7a923dd-1bed-4748-8774-c0efa6640e52-kube-api-access-pb7lx\") pod \"controller-manager-7d5766f97d-tbnns\" (UID: \"f7a923dd-1bed-4748-8774-c0efa6640e52\") " pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.369505 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.381758 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.637806 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj"] Oct 07 21:08:19 crc kubenswrapper[4926]: W1007 21:08:19.658610 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d4d7e08_b700_461a_9962_1119d0902eb6.slice/crio-e42410bf85dd2452623b999ba0b58d0ebe05d5345bfe03401fe2d8f03abb03f2 WatchSource:0}: Error finding container e42410bf85dd2452623b999ba0b58d0ebe05d5345bfe03401fe2d8f03abb03f2: Status 404 returned error can't find the container with id e42410bf85dd2452623b999ba0b58d0ebe05d5345bfe03401fe2d8f03abb03f2 Oct 07 21:08:19 crc kubenswrapper[4926]: I1007 21:08:19.739898 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d5766f97d-tbnns"] Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.559228 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" event={"ID":"f7a923dd-1bed-4748-8774-c0efa6640e52","Type":"ContainerStarted","Data":"4d4306e4f64d71f24841ee52715c1c0138b5b5fc38344e42b7e9ec9338071206"} Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.559623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" event={"ID":"f7a923dd-1bed-4748-8774-c0efa6640e52","Type":"ContainerStarted","Data":"ddee5db35544fc1b026b884379ecb87859f4fe625c08d9243c5dc71eea46e08b"} Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.559643 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.569670 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" event={"ID":"7d4d7e08-b700-461a-9962-1119d0902eb6","Type":"ContainerStarted","Data":"7f48a52b60d3ccc1da2209c3c1825fc26c73d3862c13b204fc0ef69872858717"} Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.569704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" event={"ID":"7d4d7e08-b700-461a-9962-1119d0902eb6","Type":"ContainerStarted","Data":"e42410bf85dd2452623b999ba0b58d0ebe05d5345bfe03401fe2d8f03abb03f2"} Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.570395 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.570809 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.583017 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.587811 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d5766f97d-tbnns" podStartSLOduration=3.587781578 podStartE2EDuration="3.587781578s" podCreationTimestamp="2025-10-07 21:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:08:20.58260613 +0000 UTC m=+750.621187280" watchObservedRunningTime="2025-10-07 21:08:20.587781578 +0000 UTC m=+750.626362728" Oct 07 21:08:20 crc kubenswrapper[4926]: I1007 21:08:20.634610 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86cbbcd584-2djtj" podStartSLOduration=3.634575486 podStartE2EDuration="3.634575486s" podCreationTimestamp="2025-10-07 21:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:08:20.613570576 +0000 UTC m=+750.652151726" watchObservedRunningTime="2025-10-07 21:08:20.634575486 +0000 UTC m=+750.673156636" Oct 07 21:08:25 crc kubenswrapper[4926]: I1007 21:08:25.146889 4926 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.028049 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd"] Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.029762 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.032430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.032606 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.032714 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.032850 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.032883 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-xqdpm" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.047604 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd"] Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.139231 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe0c800e-5525-4e10-8bb1-43a79192b2e8-webhook-cert\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.139360 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe0c800e-5525-4e10-8bb1-43a79192b2e8-apiservice-cert\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.139414 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j2l2\" (UniqueName: \"kubernetes.io/projected/fe0c800e-5525-4e10-8bb1-43a79192b2e8-kube-api-access-8j2l2\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.241313 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe0c800e-5525-4e10-8bb1-43a79192b2e8-apiservice-cert\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.241400 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j2l2\" (UniqueName: \"kubernetes.io/projected/fe0c800e-5525-4e10-8bb1-43a79192b2e8-kube-api-access-8j2l2\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.241444 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe0c800e-5525-4e10-8bb1-43a79192b2e8-webhook-cert\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.249228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe0c800e-5525-4e10-8bb1-43a79192b2e8-webhook-cert\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.249692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe0c800e-5525-4e10-8bb1-43a79192b2e8-apiservice-cert\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.257786 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j2l2\" (UniqueName: \"kubernetes.io/projected/fe0c800e-5525-4e10-8bb1-43a79192b2e8-kube-api-access-8j2l2\") pod \"metallb-operator-controller-manager-74978ff576-w6wnd\" (UID: \"fe0c800e-5525-4e10-8bb1-43a79192b2e8\") " pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.347394 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.391489 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8"] Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.392504 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.395096 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-q4g2h" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.395288 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.396018 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.407688 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8"] Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.445020 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/472d3ee6-7ecc-471c-842c-94afa5bda368-apiservice-cert\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.445097 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d2x8\" (UniqueName: \"kubernetes.io/projected/472d3ee6-7ecc-471c-842c-94afa5bda368-kube-api-access-6d2x8\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.445140 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/472d3ee6-7ecc-471c-842c-94afa5bda368-webhook-cert\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.546075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/472d3ee6-7ecc-471c-842c-94afa5bda368-apiservice-cert\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.546672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d2x8\" (UniqueName: \"kubernetes.io/projected/472d3ee6-7ecc-471c-842c-94afa5bda368-kube-api-access-6d2x8\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.546714 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/472d3ee6-7ecc-471c-842c-94afa5bda368-webhook-cert\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.574425 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/472d3ee6-7ecc-471c-842c-94afa5bda368-webhook-cert\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.574513 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d2x8\" (UniqueName: \"kubernetes.io/projected/472d3ee6-7ecc-471c-842c-94afa5bda368-kube-api-access-6d2x8\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.574742 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/472d3ee6-7ecc-471c-842c-94afa5bda368-apiservice-cert\") pod \"metallb-operator-webhook-server-5f56c487b-tdcp8\" (UID: \"472d3ee6-7ecc-471c-842c-94afa5bda368\") " pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.721926 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:29 crc kubenswrapper[4926]: I1007 21:08:29.826853 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd"] Oct 07 21:08:29 crc kubenswrapper[4926]: W1007 21:08:29.840521 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe0c800e_5525_4e10_8bb1_43a79192b2e8.slice/crio-ecd9aec5841df7448c7fd17698d0f27eb2fbe1d1290f6a1a09598a407ee20442 WatchSource:0}: Error finding container ecd9aec5841df7448c7fd17698d0f27eb2fbe1d1290f6a1a09598a407ee20442: Status 404 returned error can't find the container with id ecd9aec5841df7448c7fd17698d0f27eb2fbe1d1290f6a1a09598a407ee20442 Oct 07 21:08:30 crc kubenswrapper[4926]: I1007 21:08:30.134448 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8"] Oct 07 21:08:30 crc kubenswrapper[4926]: W1007 21:08:30.145513 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod472d3ee6_7ecc_471c_842c_94afa5bda368.slice/crio-8f9f0c281feffa6ea436a611d3b2ec7db68d919967ad3dab5c5c66d90ab595e4 WatchSource:0}: Error finding container 8f9f0c281feffa6ea436a611d3b2ec7db68d919967ad3dab5c5c66d90ab595e4: Status 404 returned error can't find the container with id 8f9f0c281feffa6ea436a611d3b2ec7db68d919967ad3dab5c5c66d90ab595e4 Oct 07 21:08:30 crc kubenswrapper[4926]: I1007 21:08:30.637005 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" event={"ID":"fe0c800e-5525-4e10-8bb1-43a79192b2e8","Type":"ContainerStarted","Data":"ecd9aec5841df7448c7fd17698d0f27eb2fbe1d1290f6a1a09598a407ee20442"} Oct 07 21:08:30 crc kubenswrapper[4926]: I1007 21:08:30.638967 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" event={"ID":"472d3ee6-7ecc-471c-842c-94afa5bda368","Type":"ContainerStarted","Data":"8f9f0c281feffa6ea436a611d3b2ec7db68d919967ad3dab5c5c66d90ab595e4"} Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.209143 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.209445 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.209490 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.210077 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b152fe038f21276c6a7128f942977a43dfd86caa4be6d3a00a66f622f88a2161"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.210137 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://b152fe038f21276c6a7128f942977a43dfd86caa4be6d3a00a66f622f88a2161" gracePeriod=600 Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.661350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" event={"ID":"fe0c800e-5525-4e10-8bb1-43a79192b2e8","Type":"ContainerStarted","Data":"429d09322028d4892f8d7c6b36fe4f67d71af4e9bbe516a0fd4710a84d73a217"} Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.661944 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.675009 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="b152fe038f21276c6a7128f942977a43dfd86caa4be6d3a00a66f622f88a2161" exitCode=0 Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.675071 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"b152fe038f21276c6a7128f942977a43dfd86caa4be6d3a00a66f622f88a2161"} Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.675109 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"a7195a5bc10effcad121b73904803864abd9373d3dda56955bffbf5b0e3d1427"} Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.675129 4926 scope.go:117] "RemoveContainer" containerID="ea4910515b9deac1bdd55447e1b44ec9155b5c86397c1d6d38253d2a05662255" Oct 07 21:08:33 crc kubenswrapper[4926]: I1007 21:08:33.700716 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" podStartSLOduration=1.33119339 podStartE2EDuration="4.700701528s" podCreationTimestamp="2025-10-07 21:08:29 +0000 UTC" firstStartedPulling="2025-10-07 21:08:29.843007878 +0000 UTC m=+759.881589028" lastFinishedPulling="2025-10-07 21:08:33.212516016 +0000 UTC m=+763.251097166" observedRunningTime="2025-10-07 21:08:33.696891389 +0000 UTC m=+763.735472539" watchObservedRunningTime="2025-10-07 21:08:33.700701528 +0000 UTC m=+763.739282678" Oct 07 21:08:35 crc kubenswrapper[4926]: I1007 21:08:35.706626 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" event={"ID":"472d3ee6-7ecc-471c-842c-94afa5bda368","Type":"ContainerStarted","Data":"5c3c0feeb5fdbd78c3c4396f1508203966e2dcdfadfec90b8d4a4dbd53686643"} Oct 07 21:08:35 crc kubenswrapper[4926]: I1007 21:08:35.707588 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:35 crc kubenswrapper[4926]: I1007 21:08:35.742856 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" podStartSLOduration=1.5456523739999999 podStartE2EDuration="6.742814572s" podCreationTimestamp="2025-10-07 21:08:29 +0000 UTC" firstStartedPulling="2025-10-07 21:08:30.151021177 +0000 UTC m=+760.189602327" lastFinishedPulling="2025-10-07 21:08:35.348183375 +0000 UTC m=+765.386764525" observedRunningTime="2025-10-07 21:08:35.730266963 +0000 UTC m=+765.768848123" watchObservedRunningTime="2025-10-07 21:08:35.742814572 +0000 UTC m=+765.781395762" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.319555 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xht7b"] Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.321552 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.335309 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xht7b"] Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.363558 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q82p\" (UniqueName: \"kubernetes.io/projected/295d5a55-c552-4a46-9f2a-8fae2b48f390-kube-api-access-7q82p\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.363931 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-utilities\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.364073 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-catalog-content\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.465285 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-catalog-content\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.465367 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q82p\" (UniqueName: \"kubernetes.io/projected/295d5a55-c552-4a46-9f2a-8fae2b48f390-kube-api-access-7q82p\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.465442 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-utilities\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.466122 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-catalog-content\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.466211 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-utilities\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.486477 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q82p\" (UniqueName: \"kubernetes.io/projected/295d5a55-c552-4a46-9f2a-8fae2b48f390-kube-api-access-7q82p\") pod \"certified-operators-xht7b\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:36 crc kubenswrapper[4926]: I1007 21:08:36.643554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:37 crc kubenswrapper[4926]: I1007 21:08:37.217975 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xht7b"] Oct 07 21:08:37 crc kubenswrapper[4926]: I1007 21:08:37.722895 4926 generic.go:334] "Generic (PLEG): container finished" podID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerID="c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26" exitCode=0 Oct 07 21:08:37 crc kubenswrapper[4926]: I1007 21:08:37.722975 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xht7b" event={"ID":"295d5a55-c552-4a46-9f2a-8fae2b48f390","Type":"ContainerDied","Data":"c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26"} Oct 07 21:08:37 crc kubenswrapper[4926]: I1007 21:08:37.723257 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xht7b" event={"ID":"295d5a55-c552-4a46-9f2a-8fae2b48f390","Type":"ContainerStarted","Data":"247994f2c55feb15c62781b7b792ce2f10db988d2b6b6a4152e1b5383d51d33f"} Oct 07 21:08:39 crc kubenswrapper[4926]: I1007 21:08:39.739101 4926 generic.go:334] "Generic (PLEG): container finished" podID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerID="ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7" exitCode=0 Oct 07 21:08:39 crc kubenswrapper[4926]: I1007 21:08:39.739174 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xht7b" event={"ID":"295d5a55-c552-4a46-9f2a-8fae2b48f390","Type":"ContainerDied","Data":"ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7"} Oct 07 21:08:40 crc kubenswrapper[4926]: I1007 21:08:40.747920 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xht7b" event={"ID":"295d5a55-c552-4a46-9f2a-8fae2b48f390","Type":"ContainerStarted","Data":"85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952"} Oct 07 21:08:40 crc kubenswrapper[4926]: I1007 21:08:40.770962 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xht7b" podStartSLOduration=2.288666863 podStartE2EDuration="4.770940336s" podCreationTimestamp="2025-10-07 21:08:36 +0000 UTC" firstStartedPulling="2025-10-07 21:08:37.724317123 +0000 UTC m=+767.762898273" lastFinishedPulling="2025-10-07 21:08:40.206590596 +0000 UTC m=+770.245171746" observedRunningTime="2025-10-07 21:08:40.767145548 +0000 UTC m=+770.805726698" watchObservedRunningTime="2025-10-07 21:08:40.770940336 +0000 UTC m=+770.809521496" Oct 07 21:08:46 crc kubenswrapper[4926]: I1007 21:08:46.644422 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:46 crc kubenswrapper[4926]: I1007 21:08:46.644939 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:46 crc kubenswrapper[4926]: I1007 21:08:46.692295 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:46 crc kubenswrapper[4926]: I1007 21:08:46.851263 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:46 crc kubenswrapper[4926]: I1007 21:08:46.930966 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xht7b"] Oct 07 21:08:48 crc kubenswrapper[4926]: I1007 21:08:48.806807 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xht7b" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="registry-server" containerID="cri-o://85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952" gracePeriod=2 Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.331952 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.445457 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q82p\" (UniqueName: \"kubernetes.io/projected/295d5a55-c552-4a46-9f2a-8fae2b48f390-kube-api-access-7q82p\") pod \"295d5a55-c552-4a46-9f2a-8fae2b48f390\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.445735 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-catalog-content\") pod \"295d5a55-c552-4a46-9f2a-8fae2b48f390\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.445816 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-utilities\") pod \"295d5a55-c552-4a46-9f2a-8fae2b48f390\" (UID: \"295d5a55-c552-4a46-9f2a-8fae2b48f390\") " Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.448387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-utilities" (OuterVolumeSpecName: "utilities") pod "295d5a55-c552-4a46-9f2a-8fae2b48f390" (UID: "295d5a55-c552-4a46-9f2a-8fae2b48f390"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.454969 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295d5a55-c552-4a46-9f2a-8fae2b48f390-kube-api-access-7q82p" (OuterVolumeSpecName: "kube-api-access-7q82p") pod "295d5a55-c552-4a46-9f2a-8fae2b48f390" (UID: "295d5a55-c552-4a46-9f2a-8fae2b48f390"). InnerVolumeSpecName "kube-api-access-7q82p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.543299 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "295d5a55-c552-4a46-9f2a-8fae2b48f390" (UID: "295d5a55-c552-4a46-9f2a-8fae2b48f390"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.547835 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.548104 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295d5a55-c552-4a46-9f2a-8fae2b48f390-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.548294 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q82p\" (UniqueName: \"kubernetes.io/projected/295d5a55-c552-4a46-9f2a-8fae2b48f390-kube-api-access-7q82p\") on node \"crc\" DevicePath \"\"" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.730558 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5f56c487b-tdcp8" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.820675 4926 generic.go:334] "Generic (PLEG): container finished" podID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerID="85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952" exitCode=0 Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.820731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xht7b" event={"ID":"295d5a55-c552-4a46-9f2a-8fae2b48f390","Type":"ContainerDied","Data":"85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952"} Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.820769 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xht7b" event={"ID":"295d5a55-c552-4a46-9f2a-8fae2b48f390","Type":"ContainerDied","Data":"247994f2c55feb15c62781b7b792ce2f10db988d2b6b6a4152e1b5383d51d33f"} Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.820800 4926 scope.go:117] "RemoveContainer" containerID="85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.820986 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xht7b" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.883749 4926 scope.go:117] "RemoveContainer" containerID="ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.887375 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xht7b"] Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.900640 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xht7b"] Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.927363 4926 scope.go:117] "RemoveContainer" containerID="c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.952714 4926 scope.go:117] "RemoveContainer" containerID="85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952" Oct 07 21:08:49 crc kubenswrapper[4926]: E1007 21:08:49.953290 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952\": container with ID starting with 85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952 not found: ID does not exist" containerID="85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.953324 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952"} err="failed to get container status \"85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952\": rpc error: code = NotFound desc = could not find container \"85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952\": container with ID starting with 85d476b4f5187e6a78681c54309d90af52c8b2e58ae94d8f53551aa80ec87952 not found: ID does not exist" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.953347 4926 scope.go:117] "RemoveContainer" containerID="ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7" Oct 07 21:08:49 crc kubenswrapper[4926]: E1007 21:08:49.953623 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7\": container with ID starting with ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7 not found: ID does not exist" containerID="ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.953642 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7"} err="failed to get container status \"ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7\": rpc error: code = NotFound desc = could not find container \"ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7\": container with ID starting with ef086efb36509e5dccf8df1b8ce9d0dd31e3ae28b02b11bcd338179c3d6a99d7 not found: ID does not exist" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.953653 4926 scope.go:117] "RemoveContainer" containerID="c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26" Oct 07 21:08:49 crc kubenswrapper[4926]: E1007 21:08:49.953811 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26\": container with ID starting with c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26 not found: ID does not exist" containerID="c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26" Oct 07 21:08:49 crc kubenswrapper[4926]: I1007 21:08:49.953827 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26"} err="failed to get container status \"c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26\": rpc error: code = NotFound desc = could not find container \"c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26\": container with ID starting with c8b82734f1ce8f13afdec8ecef523a98a03e3488b0ef8405baac19fbc2da4c26 not found: ID does not exist" Oct 07 21:08:50 crc kubenswrapper[4926]: I1007 21:08:50.694391 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" path="/var/lib/kubelet/pods/295d5a55-c552-4a46-9f2a-8fae2b48f390/volumes" Oct 07 21:09:09 crc kubenswrapper[4926]: I1007 21:09:09.352948 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-74978ff576-w6wnd" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.205101 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-dh249"] Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.205367 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="extract-utilities" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.205382 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="extract-utilities" Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.205415 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="extract-content" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.205423 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="extract-content" Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.205438 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="registry-server" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.205447 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="registry-server" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.205582 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="295d5a55-c552-4a46-9f2a-8fae2b48f390" containerName="registry-server" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.206056 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.208889 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.210600 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-624xv" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.216661 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-86cft"] Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.219573 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.219746 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-dh249"] Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.221124 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.222526 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-sockets\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-reloader\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243361 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c050d21-9d7e-4e41-9a8e-991580804642-cert\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243399 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc568\" (UniqueName: \"kubernetes.io/projected/5c050d21-9d7e-4e41-9a8e-991580804642-kube-api-access-zc568\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243481 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243527 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-conf\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243542 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7b6h\" (UniqueName: \"kubernetes.io/projected/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-kube-api-access-p7b6h\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243663 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-startup\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.243746 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics-certs\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.317572 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mg6kr"] Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.318624 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.320798 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.320963 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dj8x7" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.320967 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.321822 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.323572 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-xdk65"] Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.324779 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.327614 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.335898 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-xdk65"] Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345515 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-sockets\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345578 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-reloader\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345609 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c050d21-9d7e-4e41-9a8e-991580804642-cert\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345637 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc568\" (UniqueName: \"kubernetes.io/projected/5c050d21-9d7e-4e41-9a8e-991580804642-kube-api-access-zc568\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345665 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345685 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-conf\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345701 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7b6h\" (UniqueName: \"kubernetes.io/projected/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-kube-api-access-p7b6h\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345721 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-startup\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.345750 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics-certs\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.345899 4926 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.345957 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics-certs podName:d5f9b420-a8a0-4262-b6aa-49af0dc0e48c nodeName:}" failed. No retries permitted until 2025-10-07 21:09:10.84593857 +0000 UTC m=+800.884519720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics-certs") pod "frr-k8s-86cft" (UID: "d5f9b420-a8a0-4262-b6aa-49af0dc0e48c") : secret "frr-k8s-certs-secret" not found Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.346179 4926 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.346330 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c050d21-9d7e-4e41-9a8e-991580804642-cert podName:5c050d21-9d7e-4e41-9a8e-991580804642 nodeName:}" failed. No retries permitted until 2025-10-07 21:09:10.846314201 +0000 UTC m=+800.884895351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5c050d21-9d7e-4e41-9a8e-991580804642-cert") pod "frr-k8s-webhook-server-64bf5d555-dh249" (UID: "5c050d21-9d7e-4e41-9a8e-991580804642") : secret "frr-k8s-webhook-server-cert" not found Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.346346 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-reloader\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.346426 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-sockets\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.346569 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-conf\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.346718 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.346938 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-frr-startup\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.366900 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc568\" (UniqueName: \"kubernetes.io/projected/5c050d21-9d7e-4e41-9a8e-991580804642-kube-api-access-zc568\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.368914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7b6h\" (UniqueName: \"kubernetes.io/projected/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-kube-api-access-p7b6h\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447556 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17a8cd66-733c-4d61-83db-d0e56464b905-cert\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447697 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/17a8cd66-733c-4d61-83db-d0e56464b905-metrics-certs\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447761 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4phv8\" (UniqueName: \"kubernetes.io/projected/17a8cd66-733c-4d61-83db-d0e56464b905-kube-api-access-4phv8\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447783 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh6b2\" (UniqueName: \"kubernetes.io/projected/138368c1-eaa9-42ba-8b93-67bf6396ef26-kube-api-access-dh6b2\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447816 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/138368c1-eaa9-42ba-8b93-67bf6396ef26-metallb-excludel2\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.447872 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-metrics-certs\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548532 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh6b2\" (UniqueName: \"kubernetes.io/projected/138368c1-eaa9-42ba-8b93-67bf6396ef26-kube-api-access-dh6b2\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548567 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4phv8\" (UniqueName: \"kubernetes.io/projected/17a8cd66-733c-4d61-83db-d0e56464b905-kube-api-access-4phv8\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548598 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/138368c1-eaa9-42ba-8b93-67bf6396ef26-metallb-excludel2\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548638 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-metrics-certs\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17a8cd66-733c-4d61-83db-d0e56464b905-cert\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.548708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/17a8cd66-733c-4d61-83db-d0e56464b905-metrics-certs\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.549104 4926 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 21:09:10 crc kubenswrapper[4926]: E1007 21:09:10.549229 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist podName:138368c1-eaa9-42ba-8b93-67bf6396ef26 nodeName:}" failed. No retries permitted until 2025-10-07 21:09:11.049172917 +0000 UTC m=+801.087754087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist") pod "speaker-mg6kr" (UID: "138368c1-eaa9-42ba-8b93-67bf6396ef26") : secret "metallb-memberlist" not found Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.550093 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/138368c1-eaa9-42ba-8b93-67bf6396ef26-metallb-excludel2\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.551042 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.557765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/17a8cd66-733c-4d61-83db-d0e56464b905-metrics-certs\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.562107 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-metrics-certs\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.562718 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17a8cd66-733c-4d61-83db-d0e56464b905-cert\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.567331 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4phv8\" (UniqueName: \"kubernetes.io/projected/17a8cd66-733c-4d61-83db-d0e56464b905-kube-api-access-4phv8\") pod \"controller-68d546b9d8-xdk65\" (UID: \"17a8cd66-733c-4d61-83db-d0e56464b905\") " pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.581412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh6b2\" (UniqueName: \"kubernetes.io/projected/138368c1-eaa9-42ba-8b93-67bf6396ef26-kube-api-access-dh6b2\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.641051 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.853942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics-certs\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.854450 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c050d21-9d7e-4e41-9a8e-991580804642-cert\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.857662 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5f9b420-a8a0-4262-b6aa-49af0dc0e48c-metrics-certs\") pod \"frr-k8s-86cft\" (UID: \"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c\") " pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:10 crc kubenswrapper[4926]: I1007 21:09:10.859632 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c050d21-9d7e-4e41-9a8e-991580804642-cert\") pod \"frr-k8s-webhook-server-64bf5d555-dh249\" (UID: \"5c050d21-9d7e-4e41-9a8e-991580804642\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:11 crc kubenswrapper[4926]: I1007 21:09:11.057256 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:11 crc kubenswrapper[4926]: E1007 21:09:11.057459 4926 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 21:09:11 crc kubenswrapper[4926]: E1007 21:09:11.057518 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist podName:138368c1-eaa9-42ba-8b93-67bf6396ef26 nodeName:}" failed. No retries permitted until 2025-10-07 21:09:12.057500963 +0000 UTC m=+802.096082133 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist") pod "speaker-mg6kr" (UID: "138368c1-eaa9-42ba-8b93-67bf6396ef26") : secret "metallb-memberlist" not found Oct 07 21:09:11 crc kubenswrapper[4926]: I1007 21:09:11.126465 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:11 crc kubenswrapper[4926]: I1007 21:09:11.145322 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:11 crc kubenswrapper[4926]: I1007 21:09:11.174352 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-xdk65"] Oct 07 21:09:11 crc kubenswrapper[4926]: I1007 21:09:11.571288 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-dh249"] Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.005362 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"619a3f04c4d138707b750bd1f2271d74dd9991615c1c63d156cdd4e9569544aa"} Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.008217 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" event={"ID":"5c050d21-9d7e-4e41-9a8e-991580804642","Type":"ContainerStarted","Data":"10d20f6e1fc2042f2f53466157d329224b1e6156713db12e9ca8ed11850f836d"} Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.010878 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-xdk65" event={"ID":"17a8cd66-733c-4d61-83db-d0e56464b905","Type":"ContainerStarted","Data":"242cd8ddb63e6edc282df40f6df393dcee4bc0d8873f81c10fbee526b6ef59a6"} Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.010984 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-xdk65" event={"ID":"17a8cd66-733c-4d61-83db-d0e56464b905","Type":"ContainerStarted","Data":"47600fd400973a090af51e37854691be177fc93f9bdb8e74e9f777bf5e7e378f"} Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.011010 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-xdk65" event={"ID":"17a8cd66-733c-4d61-83db-d0e56464b905","Type":"ContainerStarted","Data":"bd58de6e18e66cad4f42e34d9d4fcd60ddee845d0705214eacbf90c8f55cb580"} Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.011594 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.039840 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-xdk65" podStartSLOduration=2.039805652 podStartE2EDuration="2.039805652s" podCreationTimestamp="2025-10-07 21:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:09:12.035782335 +0000 UTC m=+802.074363485" watchObservedRunningTime="2025-10-07 21:09:12.039805652 +0000 UTC m=+802.078386812" Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.074580 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.084366 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/138368c1-eaa9-42ba-8b93-67bf6396ef26-memberlist\") pod \"speaker-mg6kr\" (UID: \"138368c1-eaa9-42ba-8b93-67bf6396ef26\") " pod="metallb-system/speaker-mg6kr" Oct 07 21:09:12 crc kubenswrapper[4926]: I1007 21:09:12.135414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mg6kr" Oct 07 21:09:12 crc kubenswrapper[4926]: W1007 21:09:12.162704 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod138368c1_eaa9_42ba_8b93_67bf6396ef26.slice/crio-511ec77ccbe874a3c735c3cbd4cc7b6193f104084c30a09120f8afecb357daf7 WatchSource:0}: Error finding container 511ec77ccbe874a3c735c3cbd4cc7b6193f104084c30a09120f8afecb357daf7: Status 404 returned error can't find the container with id 511ec77ccbe874a3c735c3cbd4cc7b6193f104084c30a09120f8afecb357daf7 Oct 07 21:09:13 crc kubenswrapper[4926]: I1007 21:09:13.023788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mg6kr" event={"ID":"138368c1-eaa9-42ba-8b93-67bf6396ef26","Type":"ContainerStarted","Data":"12c2dddd3d84801919618df1a86dd9b36bf1002e43927e5c7c50bdc8f7847a3f"} Oct 07 21:09:13 crc kubenswrapper[4926]: I1007 21:09:13.024187 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mg6kr" event={"ID":"138368c1-eaa9-42ba-8b93-67bf6396ef26","Type":"ContainerStarted","Data":"c5ec3477a488242c58c23ac48b316ea947b164b6a3d994351ff50568323a7c76"} Oct 07 21:09:13 crc kubenswrapper[4926]: I1007 21:09:13.024239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mg6kr" event={"ID":"138368c1-eaa9-42ba-8b93-67bf6396ef26","Type":"ContainerStarted","Data":"511ec77ccbe874a3c735c3cbd4cc7b6193f104084c30a09120f8afecb357daf7"} Oct 07 21:09:13 crc kubenswrapper[4926]: I1007 21:09:13.024452 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mg6kr" Oct 07 21:09:13 crc kubenswrapper[4926]: I1007 21:09:13.051082 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mg6kr" podStartSLOduration=3.051053056 podStartE2EDuration="3.051053056s" podCreationTimestamp="2025-10-07 21:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:09:13.048644356 +0000 UTC m=+803.087225556" watchObservedRunningTime="2025-10-07 21:09:13.051053056 +0000 UTC m=+803.089634296" Oct 07 21:09:20 crc kubenswrapper[4926]: I1007 21:09:20.072697 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" event={"ID":"5c050d21-9d7e-4e41-9a8e-991580804642","Type":"ContainerStarted","Data":"ae8cec1f101b47f778ca8e1eb1799874363593b9d56ce348ec92a7c5fdd47b3f"} Oct 07 21:09:20 crc kubenswrapper[4926]: I1007 21:09:20.074691 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:20 crc kubenswrapper[4926]: I1007 21:09:20.075442 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5f9b420-a8a0-4262-b6aa-49af0dc0e48c" containerID="0590296e1efbb7d467c22f290ec08f1a0a7894e0795b19555200588b2c579823" exitCode=0 Oct 07 21:09:20 crc kubenswrapper[4926]: I1007 21:09:20.075528 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerDied","Data":"0590296e1efbb7d467c22f290ec08f1a0a7894e0795b19555200588b2c579823"} Oct 07 21:09:20 crc kubenswrapper[4926]: I1007 21:09:20.099658 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" podStartSLOduration=2.456787284 podStartE2EDuration="10.09962582s" podCreationTimestamp="2025-10-07 21:09:10 +0000 UTC" firstStartedPulling="2025-10-07 21:09:11.580934039 +0000 UTC m=+801.619515219" lastFinishedPulling="2025-10-07 21:09:19.223772605 +0000 UTC m=+809.262353755" observedRunningTime="2025-10-07 21:09:20.093389258 +0000 UTC m=+810.131970418" watchObservedRunningTime="2025-10-07 21:09:20.09962582 +0000 UTC m=+810.138207000" Oct 07 21:09:21 crc kubenswrapper[4926]: I1007 21:09:21.089575 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5f9b420-a8a0-4262-b6aa-49af0dc0e48c" containerID="b4a6e6b1cc0e1cabdb721de6ccc4ebf2631f7de387ccc32b34cebe3dbda7459b" exitCode=0 Oct 07 21:09:21 crc kubenswrapper[4926]: I1007 21:09:21.089669 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerDied","Data":"b4a6e6b1cc0e1cabdb721de6ccc4ebf2631f7de387ccc32b34cebe3dbda7459b"} Oct 07 21:09:22 crc kubenswrapper[4926]: I1007 21:09:22.109653 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5f9b420-a8a0-4262-b6aa-49af0dc0e48c" containerID="7a706c02e34af6754f6339fa46e203b1b6544193afab85110aba411fc3bfcf3f" exitCode=0 Oct 07 21:09:22 crc kubenswrapper[4926]: I1007 21:09:22.109721 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerDied","Data":"7a706c02e34af6754f6339fa46e203b1b6544193afab85110aba411fc3bfcf3f"} Oct 07 21:09:22 crc kubenswrapper[4926]: I1007 21:09:22.147331 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mg6kr" Oct 07 21:09:23 crc kubenswrapper[4926]: I1007 21:09:23.127991 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"cc50bb2be8074ff23991a8922e40ae7c9ab8595627bdd3710f02edf5a4d863f9"} Oct 07 21:09:23 crc kubenswrapper[4926]: I1007 21:09:23.128373 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"84718deacc70d71c99447ab4d8496f15cc9eafa5cf9082adf8897151b8d582d7"} Oct 07 21:09:23 crc kubenswrapper[4926]: I1007 21:09:23.128386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"06d37dad5650eea60281b618c20f66e34e500fcfe916f61da13f4560997da6b6"} Oct 07 21:09:23 crc kubenswrapper[4926]: I1007 21:09:23.128395 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"feb722531ea26424164104457aea2b73b171b9689d0e6ff9a0175cf782b765d5"} Oct 07 21:09:23 crc kubenswrapper[4926]: I1007 21:09:23.128405 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"e385b6f51a5cb85effc938178aa9896c8030053b3a09e8877b4dfcadf6f525bf"} Oct 07 21:09:24 crc kubenswrapper[4926]: I1007 21:09:24.140487 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-86cft" event={"ID":"d5f9b420-a8a0-4262-b6aa-49af0dc0e48c","Type":"ContainerStarted","Data":"d60925a115966675db73199ab7d1e57c9e7a0bbe2f3bd1807e661089742750ed"} Oct 07 21:09:24 crc kubenswrapper[4926]: I1007 21:09:24.141234 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:24 crc kubenswrapper[4926]: I1007 21:09:24.175622 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-86cft" podStartSLOduration=6.197607846 podStartE2EDuration="14.175603297s" podCreationTimestamp="2025-10-07 21:09:10 +0000 UTC" firstStartedPulling="2025-10-07 21:09:11.282468474 +0000 UTC m=+801.321049624" lastFinishedPulling="2025-10-07 21:09:19.260463915 +0000 UTC m=+809.299045075" observedRunningTime="2025-10-07 21:09:24.172804356 +0000 UTC m=+814.211385496" watchObservedRunningTime="2025-10-07 21:09:24.175603297 +0000 UTC m=+814.214184447" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.358374 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-bf2kz"] Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.359701 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.365411 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.365461 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fwkd2" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.366259 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.370946 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bf2kz"] Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.511030 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45mb7\" (UniqueName: \"kubernetes.io/projected/7b1c701e-d87c-4bec-8b94-8708bb356171-kube-api-access-45mb7\") pod \"openstack-operator-index-bf2kz\" (UID: \"7b1c701e-d87c-4bec-8b94-8708bb356171\") " pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.613342 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45mb7\" (UniqueName: \"kubernetes.io/projected/7b1c701e-d87c-4bec-8b94-8708bb356171-kube-api-access-45mb7\") pod \"openstack-operator-index-bf2kz\" (UID: \"7b1c701e-d87c-4bec-8b94-8708bb356171\") " pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.641244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45mb7\" (UniqueName: \"kubernetes.io/projected/7b1c701e-d87c-4bec-8b94-8708bb356171-kube-api-access-45mb7\") pod \"openstack-operator-index-bf2kz\" (UID: \"7b1c701e-d87c-4bec-8b94-8708bb356171\") " pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:25 crc kubenswrapper[4926]: I1007 21:09:25.687651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:26 crc kubenswrapper[4926]: I1007 21:09:26.145518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:26 crc kubenswrapper[4926]: I1007 21:09:26.182176 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bf2kz"] Oct 07 21:09:26 crc kubenswrapper[4926]: I1007 21:09:26.187707 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:26 crc kubenswrapper[4926]: W1007 21:09:26.188847 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b1c701e_d87c_4bec_8b94_8708bb356171.slice/crio-5fb5eb2cf73844e8ae129a86ad722d0f1b7452e2bfe861f7da981fc89dda6b8a WatchSource:0}: Error finding container 5fb5eb2cf73844e8ae129a86ad722d0f1b7452e2bfe861f7da981fc89dda6b8a: Status 404 returned error can't find the container with id 5fb5eb2cf73844e8ae129a86ad722d0f1b7452e2bfe861f7da981fc89dda6b8a Oct 07 21:09:27 crc kubenswrapper[4926]: I1007 21:09:27.167356 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bf2kz" event={"ID":"7b1c701e-d87c-4bec-8b94-8708bb356171","Type":"ContainerStarted","Data":"5fb5eb2cf73844e8ae129a86ad722d0f1b7452e2bfe861f7da981fc89dda6b8a"} Oct 07 21:09:28 crc kubenswrapper[4926]: I1007 21:09:28.926834 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bf2kz"] Oct 07 21:09:29 crc kubenswrapper[4926]: I1007 21:09:29.733493 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hbjzj"] Oct 07 21:09:29 crc kubenswrapper[4926]: I1007 21:09:29.734871 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:29 crc kubenswrapper[4926]: I1007 21:09:29.787533 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hbjzj"] Oct 07 21:09:29 crc kubenswrapper[4926]: I1007 21:09:29.883396 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzqf8\" (UniqueName: \"kubernetes.io/projected/240049fd-6fcf-493a-933e-c64ebca105b3-kube-api-access-xzqf8\") pod \"openstack-operator-index-hbjzj\" (UID: \"240049fd-6fcf-493a-933e-c64ebca105b3\") " pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:29 crc kubenswrapper[4926]: I1007 21:09:29.984724 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzqf8\" (UniqueName: \"kubernetes.io/projected/240049fd-6fcf-493a-933e-c64ebca105b3-kube-api-access-xzqf8\") pod \"openstack-operator-index-hbjzj\" (UID: \"240049fd-6fcf-493a-933e-c64ebca105b3\") " pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.010329 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzqf8\" (UniqueName: \"kubernetes.io/projected/240049fd-6fcf-493a-933e-c64ebca105b3-kube-api-access-xzqf8\") pod \"openstack-operator-index-hbjzj\" (UID: \"240049fd-6fcf-493a-933e-c64ebca105b3\") " pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.112067 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.193075 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bf2kz" event={"ID":"7b1c701e-d87c-4bec-8b94-8708bb356171","Type":"ContainerStarted","Data":"e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4"} Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.193221 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-bf2kz" podUID="7b1c701e-d87c-4bec-8b94-8708bb356171" containerName="registry-server" containerID="cri-o://e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4" gracePeriod=2 Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.213241 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-bf2kz" podStartSLOduration=1.617757874 podStartE2EDuration="5.213176785s" podCreationTimestamp="2025-10-07 21:09:25 +0000 UTC" firstStartedPulling="2025-10-07 21:09:26.191084449 +0000 UTC m=+816.229665599" lastFinishedPulling="2025-10-07 21:09:29.78650334 +0000 UTC m=+819.825084510" observedRunningTime="2025-10-07 21:09:30.209615031 +0000 UTC m=+820.248196191" watchObservedRunningTime="2025-10-07 21:09:30.213176785 +0000 UTC m=+820.251757945" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.544334 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.593131 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hbjzj"] Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.647040 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-xdk65" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.694956 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45mb7\" (UniqueName: \"kubernetes.io/projected/7b1c701e-d87c-4bec-8b94-8708bb356171-kube-api-access-45mb7\") pod \"7b1c701e-d87c-4bec-8b94-8708bb356171\" (UID: \"7b1c701e-d87c-4bec-8b94-8708bb356171\") " Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.700439 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b1c701e-d87c-4bec-8b94-8708bb356171-kube-api-access-45mb7" (OuterVolumeSpecName: "kube-api-access-45mb7") pod "7b1c701e-d87c-4bec-8b94-8708bb356171" (UID: "7b1c701e-d87c-4bec-8b94-8708bb356171"). InnerVolumeSpecName "kube-api-access-45mb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:09:30 crc kubenswrapper[4926]: I1007 21:09:30.798003 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45mb7\" (UniqueName: \"kubernetes.io/projected/7b1c701e-d87c-4bec-8b94-8708bb356171-kube-api-access-45mb7\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.134841 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-dh249" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.218103 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hbjzj" event={"ID":"240049fd-6fcf-493a-933e-c64ebca105b3","Type":"ContainerStarted","Data":"94f667842d691d299ea69857a05d4ed3cc0446a2376b39dd3988e2ba5b9b0a75"} Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.218439 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hbjzj" event={"ID":"240049fd-6fcf-493a-933e-c64ebca105b3","Type":"ContainerStarted","Data":"47e008b1cf34d807b8633fa96fb6cb8844eff2b0727fe580604ce068abd109bb"} Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.222653 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b1c701e-d87c-4bec-8b94-8708bb356171" containerID="e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4" exitCode=0 Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.222683 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bf2kz" event={"ID":"7b1c701e-d87c-4bec-8b94-8708bb356171","Type":"ContainerDied","Data":"e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4"} Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.222703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bf2kz" event={"ID":"7b1c701e-d87c-4bec-8b94-8708bb356171","Type":"ContainerDied","Data":"5fb5eb2cf73844e8ae129a86ad722d0f1b7452e2bfe861f7da981fc89dda6b8a"} Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.222723 4926 scope.go:117] "RemoveContainer" containerID="e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.222836 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bf2kz" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.233705 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hbjzj" podStartSLOduration=2.166152029 podStartE2EDuration="2.233688719s" podCreationTimestamp="2025-10-07 21:09:29 +0000 UTC" firstStartedPulling="2025-10-07 21:09:30.60697859 +0000 UTC m=+820.645559740" lastFinishedPulling="2025-10-07 21:09:30.67451527 +0000 UTC m=+820.713096430" observedRunningTime="2025-10-07 21:09:31.233490733 +0000 UTC m=+821.272071913" watchObservedRunningTime="2025-10-07 21:09:31.233688719 +0000 UTC m=+821.272269869" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.261497 4926 scope.go:117] "RemoveContainer" containerID="e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4" Oct 07 21:09:31 crc kubenswrapper[4926]: E1007 21:09:31.262701 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4\": container with ID starting with e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4 not found: ID does not exist" containerID="e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.262741 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4"} err="failed to get container status \"e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4\": rpc error: code = NotFound desc = could not find container \"e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4\": container with ID starting with e9d5239ee3a177d1cdbb988188afb9551afb83f1a8013e72723a56a418eb01e4 not found: ID does not exist" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.265462 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bf2kz"] Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.270814 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-bf2kz"] Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.532663 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ltvf2"] Oct 07 21:09:31 crc kubenswrapper[4926]: E1007 21:09:31.532896 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b1c701e-d87c-4bec-8b94-8708bb356171" containerName="registry-server" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.532906 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b1c701e-d87c-4bec-8b94-8708bb356171" containerName="registry-server" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.533022 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b1c701e-d87c-4bec-8b94-8708bb356171" containerName="registry-server" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.533808 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.556223 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ltvf2"] Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.714303 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfdzq\" (UniqueName: \"kubernetes.io/projected/4b077eed-9113-4db8-9ba9-8b5d960ad872-kube-api-access-pfdzq\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.715534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-utilities\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.715708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-catalog-content\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.819352 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfdzq\" (UniqueName: \"kubernetes.io/projected/4b077eed-9113-4db8-9ba9-8b5d960ad872-kube-api-access-pfdzq\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.819453 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-utilities\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.819555 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-catalog-content\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.820473 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-utilities\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.820565 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-catalog-content\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:31 crc kubenswrapper[4926]: I1007 21:09:31.864445 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfdzq\" (UniqueName: \"kubernetes.io/projected/4b077eed-9113-4db8-9ba9-8b5d960ad872-kube-api-access-pfdzq\") pod \"redhat-operators-ltvf2\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:32 crc kubenswrapper[4926]: I1007 21:09:32.162405 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:32 crc kubenswrapper[4926]: I1007 21:09:32.651491 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ltvf2"] Oct 07 21:09:32 crc kubenswrapper[4926]: W1007 21:09:32.662175 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b077eed_9113_4db8_9ba9_8b5d960ad872.slice/crio-da2279b34a21db6da5bd5d05a4788c7648729cfc9905e8a5a8aa9b4710680c98 WatchSource:0}: Error finding container da2279b34a21db6da5bd5d05a4788c7648729cfc9905e8a5a8aa9b4710680c98: Status 404 returned error can't find the container with id da2279b34a21db6da5bd5d05a4788c7648729cfc9905e8a5a8aa9b4710680c98 Oct 07 21:09:32 crc kubenswrapper[4926]: I1007 21:09:32.687777 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b1c701e-d87c-4bec-8b94-8708bb356171" path="/var/lib/kubelet/pods/7b1c701e-d87c-4bec-8b94-8708bb356171/volumes" Oct 07 21:09:33 crc kubenswrapper[4926]: I1007 21:09:33.244001 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerStarted","Data":"da2279b34a21db6da5bd5d05a4788c7648729cfc9905e8a5a8aa9b4710680c98"} Oct 07 21:09:34 crc kubenswrapper[4926]: I1007 21:09:34.253569 4926 generic.go:334] "Generic (PLEG): container finished" podID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerID="cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba" exitCode=0 Oct 07 21:09:34 crc kubenswrapper[4926]: I1007 21:09:34.253701 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerDied","Data":"cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba"} Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.145803 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d99m2"] Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.149628 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.162654 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99m2"] Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.262721 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerStarted","Data":"002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0"} Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.269734 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsmk6\" (UniqueName: \"kubernetes.io/projected/5ffc3969-075d-4e75-b169-2ead83ab958d-kube-api-access-gsmk6\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.269835 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-catalog-content\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.269980 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-utilities\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.371527 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-utilities\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.371627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsmk6\" (UniqueName: \"kubernetes.io/projected/5ffc3969-075d-4e75-b169-2ead83ab958d-kube-api-access-gsmk6\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.371678 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-catalog-content\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.372245 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-utilities\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.372442 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-catalog-content\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.395516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsmk6\" (UniqueName: \"kubernetes.io/projected/5ffc3969-075d-4e75-b169-2ead83ab958d-kube-api-access-gsmk6\") pod \"redhat-marketplace-d99m2\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.483595 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:35 crc kubenswrapper[4926]: I1007 21:09:35.915818 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99m2"] Oct 07 21:09:35 crc kubenswrapper[4926]: W1007 21:09:35.922109 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ffc3969_075d_4e75_b169_2ead83ab958d.slice/crio-8b3e4e530fc80b70f0f3043aaa92067c289c07e4f0bd7550b0d8a06fa6ac1234 WatchSource:0}: Error finding container 8b3e4e530fc80b70f0f3043aaa92067c289c07e4f0bd7550b0d8a06fa6ac1234: Status 404 returned error can't find the container with id 8b3e4e530fc80b70f0f3043aaa92067c289c07e4f0bd7550b0d8a06fa6ac1234 Oct 07 21:09:36 crc kubenswrapper[4926]: I1007 21:09:36.272836 4926 generic.go:334] "Generic (PLEG): container finished" podID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerID="002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0" exitCode=0 Oct 07 21:09:36 crc kubenswrapper[4926]: I1007 21:09:36.272897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerDied","Data":"002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0"} Oct 07 21:09:36 crc kubenswrapper[4926]: I1007 21:09:36.277859 4926 generic.go:334] "Generic (PLEG): container finished" podID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerID="0ce2354838ec64c67d4b36c4ec159f4fe6d49b29b6362a36e2ee310023a5209a" exitCode=0 Oct 07 21:09:36 crc kubenswrapper[4926]: I1007 21:09:36.277899 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99m2" event={"ID":"5ffc3969-075d-4e75-b169-2ead83ab958d","Type":"ContainerDied","Data":"0ce2354838ec64c67d4b36c4ec159f4fe6d49b29b6362a36e2ee310023a5209a"} Oct 07 21:09:36 crc kubenswrapper[4926]: I1007 21:09:36.277986 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99m2" event={"ID":"5ffc3969-075d-4e75-b169-2ead83ab958d","Type":"ContainerStarted","Data":"8b3e4e530fc80b70f0f3043aaa92067c289c07e4f0bd7550b0d8a06fa6ac1234"} Oct 07 21:09:37 crc kubenswrapper[4926]: I1007 21:09:37.290383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerStarted","Data":"3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608"} Oct 07 21:09:37 crc kubenswrapper[4926]: I1007 21:09:37.314601 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ltvf2" podStartSLOduration=3.792936325 podStartE2EDuration="6.314557228s" podCreationTimestamp="2025-10-07 21:09:31 +0000 UTC" firstStartedPulling="2025-10-07 21:09:34.256589362 +0000 UTC m=+824.295170542" lastFinishedPulling="2025-10-07 21:09:36.778210295 +0000 UTC m=+826.816791445" observedRunningTime="2025-10-07 21:09:37.311567431 +0000 UTC m=+827.350148601" watchObservedRunningTime="2025-10-07 21:09:37.314557228 +0000 UTC m=+827.353138378" Oct 07 21:09:38 crc kubenswrapper[4926]: I1007 21:09:38.302652 4926 generic.go:334] "Generic (PLEG): container finished" podID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerID="a5123e62d1110eced393aa10d4faab9fc1865a887a109028b6fdccb2e4ddfa3e" exitCode=0 Oct 07 21:09:38 crc kubenswrapper[4926]: I1007 21:09:38.302741 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99m2" event={"ID":"5ffc3969-075d-4e75-b169-2ead83ab958d","Type":"ContainerDied","Data":"a5123e62d1110eced393aa10d4faab9fc1865a887a109028b6fdccb2e4ddfa3e"} Oct 07 21:09:39 crc kubenswrapper[4926]: I1007 21:09:39.315394 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99m2" event={"ID":"5ffc3969-075d-4e75-b169-2ead83ab958d","Type":"ContainerStarted","Data":"360c708b0e88c28f928500e84f376b6447bafb90d15cbaa70effd78c841c29e1"} Oct 07 21:09:39 crc kubenswrapper[4926]: I1007 21:09:39.337602 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d99m2" podStartSLOduration=1.881185857 podStartE2EDuration="4.33758007s" podCreationTimestamp="2025-10-07 21:09:35 +0000 UTC" firstStartedPulling="2025-10-07 21:09:36.279401717 +0000 UTC m=+826.317982867" lastFinishedPulling="2025-10-07 21:09:38.73579592 +0000 UTC m=+828.774377080" observedRunningTime="2025-10-07 21:09:39.336602242 +0000 UTC m=+829.375183422" watchObservedRunningTime="2025-10-07 21:09:39.33758007 +0000 UTC m=+829.376161250" Oct 07 21:09:40 crc kubenswrapper[4926]: I1007 21:09:40.113168 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:40 crc kubenswrapper[4926]: I1007 21:09:40.113632 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:40 crc kubenswrapper[4926]: I1007 21:09:40.144015 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:40 crc kubenswrapper[4926]: I1007 21:09:40.350549 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hbjzj" Oct 07 21:09:41 crc kubenswrapper[4926]: I1007 21:09:41.151895 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-86cft" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.163381 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.163449 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.164969 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2"] Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.166093 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.168019 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-m2mgs" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.179771 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2"] Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.188293 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-bundle\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.188371 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stmf9\" (UniqueName: \"kubernetes.io/projected/2ab95c76-4619-40f4-a661-36f224cecca8-kube-api-access-stmf9\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.188406 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-util\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.289283 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmf9\" (UniqueName: \"kubernetes.io/projected/2ab95c76-4619-40f4-a661-36f224cecca8-kube-api-access-stmf9\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.289733 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-util\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.289882 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-bundle\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.290175 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-util\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.290522 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-bundle\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.313406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stmf9\" (UniqueName: \"kubernetes.io/projected/2ab95c76-4619-40f4-a661-36f224cecca8-kube-api-access-stmf9\") pod \"62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.485871 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:42 crc kubenswrapper[4926]: I1007 21:09:42.938369 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2"] Oct 07 21:09:43 crc kubenswrapper[4926]: I1007 21:09:43.220776 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ltvf2" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="registry-server" probeResult="failure" output=< Oct 07 21:09:43 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Oct 07 21:09:43 crc kubenswrapper[4926]: > Oct 07 21:09:43 crc kubenswrapper[4926]: I1007 21:09:43.362518 4926 generic.go:334] "Generic (PLEG): container finished" podID="2ab95c76-4619-40f4-a661-36f224cecca8" containerID="5f878d7649af402998641615ba1bb3080ec5ad843203adbca31ef6b89773ba6c" exitCode=0 Oct 07 21:09:43 crc kubenswrapper[4926]: I1007 21:09:43.362569 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" event={"ID":"2ab95c76-4619-40f4-a661-36f224cecca8","Type":"ContainerDied","Data":"5f878d7649af402998641615ba1bb3080ec5ad843203adbca31ef6b89773ba6c"} Oct 07 21:09:43 crc kubenswrapper[4926]: I1007 21:09:43.362596 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" event={"ID":"2ab95c76-4619-40f4-a661-36f224cecca8","Type":"ContainerStarted","Data":"ecb88136566d4065021c9b44a8c494c5116ee5cb471ee99385c82ded3b3520de"} Oct 07 21:09:44 crc kubenswrapper[4926]: I1007 21:09:44.376804 4926 generic.go:334] "Generic (PLEG): container finished" podID="2ab95c76-4619-40f4-a661-36f224cecca8" containerID="3d437befe23b111e49c22b728cd013251a5004f0b5b182defdeb46ec63c46d57" exitCode=0 Oct 07 21:09:44 crc kubenswrapper[4926]: I1007 21:09:44.376905 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" event={"ID":"2ab95c76-4619-40f4-a661-36f224cecca8","Type":"ContainerDied","Data":"3d437befe23b111e49c22b728cd013251a5004f0b5b182defdeb46ec63c46d57"} Oct 07 21:09:45 crc kubenswrapper[4926]: I1007 21:09:45.389363 4926 generic.go:334] "Generic (PLEG): container finished" podID="2ab95c76-4619-40f4-a661-36f224cecca8" containerID="7104851e4c784a5900b4f0f0007ea6cb4dd294b84b2566f11e181442ec429187" exitCode=0 Oct 07 21:09:45 crc kubenswrapper[4926]: I1007 21:09:45.389455 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" event={"ID":"2ab95c76-4619-40f4-a661-36f224cecca8","Type":"ContainerDied","Data":"7104851e4c784a5900b4f0f0007ea6cb4dd294b84b2566f11e181442ec429187"} Oct 07 21:09:45 crc kubenswrapper[4926]: I1007 21:09:45.483824 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:45 crc kubenswrapper[4926]: I1007 21:09:45.483897 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:45 crc kubenswrapper[4926]: I1007 21:09:45.558177 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.470616 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.768703 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.961886 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-bundle\") pod \"2ab95c76-4619-40f4-a661-36f224cecca8\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.962080 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-util\") pod \"2ab95c76-4619-40f4-a661-36f224cecca8\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.962185 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stmf9\" (UniqueName: \"kubernetes.io/projected/2ab95c76-4619-40f4-a661-36f224cecca8-kube-api-access-stmf9\") pod \"2ab95c76-4619-40f4-a661-36f224cecca8\" (UID: \"2ab95c76-4619-40f4-a661-36f224cecca8\") " Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.963175 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-bundle" (OuterVolumeSpecName: "bundle") pod "2ab95c76-4619-40f4-a661-36f224cecca8" (UID: "2ab95c76-4619-40f4-a661-36f224cecca8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.969461 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ab95c76-4619-40f4-a661-36f224cecca8-kube-api-access-stmf9" (OuterVolumeSpecName: "kube-api-access-stmf9") pod "2ab95c76-4619-40f4-a661-36f224cecca8" (UID: "2ab95c76-4619-40f4-a661-36f224cecca8"). InnerVolumeSpecName "kube-api-access-stmf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:09:46 crc kubenswrapper[4926]: I1007 21:09:46.984451 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-util" (OuterVolumeSpecName: "util") pod "2ab95c76-4619-40f4-a661-36f224cecca8" (UID: "2ab95c76-4619-40f4-a661-36f224cecca8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:09:47 crc kubenswrapper[4926]: I1007 21:09:47.064189 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-util\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:47 crc kubenswrapper[4926]: I1007 21:09:47.064282 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stmf9\" (UniqueName: \"kubernetes.io/projected/2ab95c76-4619-40f4-a661-36f224cecca8-kube-api-access-stmf9\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:47 crc kubenswrapper[4926]: I1007 21:09:47.064304 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ab95c76-4619-40f4-a661-36f224cecca8-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:47 crc kubenswrapper[4926]: I1007 21:09:47.411089 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" Oct 07 21:09:47 crc kubenswrapper[4926]: I1007 21:09:47.411089 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2" event={"ID":"2ab95c76-4619-40f4-a661-36f224cecca8","Type":"ContainerDied","Data":"ecb88136566d4065021c9b44a8c494c5116ee5cb471ee99385c82ded3b3520de"} Oct 07 21:09:47 crc kubenswrapper[4926]: I1007 21:09:47.411164 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecb88136566d4065021c9b44a8c494c5116ee5cb471ee99385c82ded3b3520de" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.123095 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99m2"] Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.125669 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d99m2" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="registry-server" containerID="cri-o://360c708b0e88c28f928500e84f376b6447bafb90d15cbaa70effd78c841c29e1" gracePeriod=2 Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.438495 4926 generic.go:334] "Generic (PLEG): container finished" podID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerID="360c708b0e88c28f928500e84f376b6447bafb90d15cbaa70effd78c841c29e1" exitCode=0 Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.438552 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99m2" event={"ID":"5ffc3969-075d-4e75-b169-2ead83ab958d","Type":"ContainerDied","Data":"360c708b0e88c28f928500e84f376b6447bafb90d15cbaa70effd78c841c29e1"} Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.532222 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2vnrc"] Oct 07 21:09:50 crc kubenswrapper[4926]: E1007 21:09:50.532524 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="extract" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.532546 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="extract" Oct 07 21:09:50 crc kubenswrapper[4926]: E1007 21:09:50.532560 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="pull" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.532569 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="pull" Oct 07 21:09:50 crc kubenswrapper[4926]: E1007 21:09:50.532592 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="util" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.532600 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="util" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.532736 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab95c76-4619-40f4-a661-36f224cecca8" containerName="extract" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.533887 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.543902 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2vnrc"] Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.560891 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.655065 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8578696b-282fd"] Oct 07 21:09:50 crc kubenswrapper[4926]: E1007 21:09:50.655380 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="registry-server" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.655403 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="registry-server" Oct 07 21:09:50 crc kubenswrapper[4926]: E1007 21:09:50.655422 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="extract-content" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.655431 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="extract-content" Oct 07 21:09:50 crc kubenswrapper[4926]: E1007 21:09:50.655456 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="extract-utilities" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.655466 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="extract-utilities" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.655600 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" containerName="registry-server" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.656444 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.659774 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-g88p5" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.691460 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8578696b-282fd"] Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.713534 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-catalog-content\") pod \"5ffc3969-075d-4e75-b169-2ead83ab958d\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.713630 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsmk6\" (UniqueName: \"kubernetes.io/projected/5ffc3969-075d-4e75-b169-2ead83ab958d-kube-api-access-gsmk6\") pod \"5ffc3969-075d-4e75-b169-2ead83ab958d\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.713659 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-utilities\") pod \"5ffc3969-075d-4e75-b169-2ead83ab958d\" (UID: \"5ffc3969-075d-4e75-b169-2ead83ab958d\") " Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.713783 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-utilities\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.713850 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-catalog-content\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.713876 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh72n\" (UniqueName: \"kubernetes.io/projected/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-kube-api-access-rh72n\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.714932 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-utilities" (OuterVolumeSpecName: "utilities") pod "5ffc3969-075d-4e75-b169-2ead83ab958d" (UID: "5ffc3969-075d-4e75-b169-2ead83ab958d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.719269 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ffc3969-075d-4e75-b169-2ead83ab958d-kube-api-access-gsmk6" (OuterVolumeSpecName: "kube-api-access-gsmk6") pod "5ffc3969-075d-4e75-b169-2ead83ab958d" (UID: "5ffc3969-075d-4e75-b169-2ead83ab958d"). InnerVolumeSpecName "kube-api-access-gsmk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.728040 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ffc3969-075d-4e75-b169-2ead83ab958d" (UID: "5ffc3969-075d-4e75-b169-2ead83ab958d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815556 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-catalog-content\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815628 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh72n\" (UniqueName: \"kubernetes.io/projected/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-kube-api-access-rh72n\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-utilities\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815749 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wg9v\" (UniqueName: \"kubernetes.io/projected/37e8795d-6050-4b61-9282-f3af17a926c8-kube-api-access-7wg9v\") pod \"openstack-operator-controller-operator-8578696b-282fd\" (UID: \"37e8795d-6050-4b61-9282-f3af17a926c8\") " pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815868 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsmk6\" (UniqueName: \"kubernetes.io/projected/5ffc3969-075d-4e75-b169-2ead83ab958d-kube-api-access-gsmk6\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815879 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.815889 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ffc3969-075d-4e75-b169-2ead83ab958d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.816722 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-catalog-content\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.816777 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-utilities\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.834262 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh72n\" (UniqueName: \"kubernetes.io/projected/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-kube-api-access-rh72n\") pod \"community-operators-2vnrc\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.870187 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.917060 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wg9v\" (UniqueName: \"kubernetes.io/projected/37e8795d-6050-4b61-9282-f3af17a926c8-kube-api-access-7wg9v\") pod \"openstack-operator-controller-operator-8578696b-282fd\" (UID: \"37e8795d-6050-4b61-9282-f3af17a926c8\") " pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.933966 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wg9v\" (UniqueName: \"kubernetes.io/projected/37e8795d-6050-4b61-9282-f3af17a926c8-kube-api-access-7wg9v\") pod \"openstack-operator-controller-operator-8578696b-282fd\" (UID: \"37e8795d-6050-4b61-9282-f3af17a926c8\") " pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:09:50 crc kubenswrapper[4926]: I1007 21:09:50.983351 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.127271 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2vnrc"] Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.446973 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99m2" event={"ID":"5ffc3969-075d-4e75-b169-2ead83ab958d","Type":"ContainerDied","Data":"8b3e4e530fc80b70f0f3043aaa92067c289c07e4f0bd7550b0d8a06fa6ac1234"} Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.447035 4926 scope.go:117] "RemoveContainer" containerID="360c708b0e88c28f928500e84f376b6447bafb90d15cbaa70effd78c841c29e1" Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.447043 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99m2" Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.448881 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerID="742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754" exitCode=0 Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.448912 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vnrc" event={"ID":"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c","Type":"ContainerDied","Data":"742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754"} Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.448927 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vnrc" event={"ID":"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c","Type":"ContainerStarted","Data":"c07ce3cc7ccdbb41fb5fa66fa29cbd2333fe7ebfa9e21a5c6413d5c83bc3f241"} Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.463550 4926 scope.go:117] "RemoveContainer" containerID="a5123e62d1110eced393aa10d4faab9fc1865a887a109028b6fdccb2e4ddfa3e" Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.486625 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8578696b-282fd"] Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.490241 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99m2"] Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.491405 4926 scope.go:117] "RemoveContainer" containerID="0ce2354838ec64c67d4b36c4ec159f4fe6d49b29b6362a36e2ee310023a5209a" Oct 07 21:09:51 crc kubenswrapper[4926]: I1007 21:09:51.495928 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99m2"] Oct 07 21:09:52 crc kubenswrapper[4926]: I1007 21:09:52.204641 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:52 crc kubenswrapper[4926]: I1007 21:09:52.246660 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:52 crc kubenswrapper[4926]: I1007 21:09:52.457871 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" event={"ID":"37e8795d-6050-4b61-9282-f3af17a926c8","Type":"ContainerStarted","Data":"d021f2ab23f076fc2629fe92233255f46077d7032c0918a082a5af09e7e6f1be"} Oct 07 21:09:52 crc kubenswrapper[4926]: I1007 21:09:52.695597 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ffc3969-075d-4e75-b169-2ead83ab958d" path="/var/lib/kubelet/pods/5ffc3969-075d-4e75-b169-2ead83ab958d/volumes" Oct 07 21:09:55 crc kubenswrapper[4926]: I1007 21:09:55.484499 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerID="60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae" exitCode=0 Oct 07 21:09:55 crc kubenswrapper[4926]: I1007 21:09:55.484590 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vnrc" event={"ID":"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c","Type":"ContainerDied","Data":"60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae"} Oct 07 21:09:55 crc kubenswrapper[4926]: I1007 21:09:55.489255 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" event={"ID":"37e8795d-6050-4b61-9282-f3af17a926c8","Type":"ContainerStarted","Data":"57a2242fe30d8ef7b5d6e62266208e5870ccaa8b630f83b53ec18c3ec15dd909"} Oct 07 21:09:56 crc kubenswrapper[4926]: I1007 21:09:56.535309 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ltvf2"] Oct 07 21:09:56 crc kubenswrapper[4926]: I1007 21:09:56.535655 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ltvf2" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="registry-server" containerID="cri-o://3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608" gracePeriod=2 Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.291335 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.316667 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-utilities\") pod \"4b077eed-9113-4db8-9ba9-8b5d960ad872\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.316794 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-catalog-content\") pod \"4b077eed-9113-4db8-9ba9-8b5d960ad872\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.316819 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfdzq\" (UniqueName: \"kubernetes.io/projected/4b077eed-9113-4db8-9ba9-8b5d960ad872-kube-api-access-pfdzq\") pod \"4b077eed-9113-4db8-9ba9-8b5d960ad872\" (UID: \"4b077eed-9113-4db8-9ba9-8b5d960ad872\") " Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.317771 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-utilities" (OuterVolumeSpecName: "utilities") pod "4b077eed-9113-4db8-9ba9-8b5d960ad872" (UID: "4b077eed-9113-4db8-9ba9-8b5d960ad872"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.318239 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.364275 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b077eed-9113-4db8-9ba9-8b5d960ad872-kube-api-access-pfdzq" (OuterVolumeSpecName: "kube-api-access-pfdzq") pod "4b077eed-9113-4db8-9ba9-8b5d960ad872" (UID: "4b077eed-9113-4db8-9ba9-8b5d960ad872"). InnerVolumeSpecName "kube-api-access-pfdzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.412666 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b077eed-9113-4db8-9ba9-8b5d960ad872" (UID: "4b077eed-9113-4db8-9ba9-8b5d960ad872"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.418727 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b077eed-9113-4db8-9ba9-8b5d960ad872-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.418760 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfdzq\" (UniqueName: \"kubernetes.io/projected/4b077eed-9113-4db8-9ba9-8b5d960ad872-kube-api-access-pfdzq\") on node \"crc\" DevicePath \"\"" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.508177 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vnrc" event={"ID":"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c","Type":"ContainerStarted","Data":"ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146"} Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.513212 4926 generic.go:334] "Generic (PLEG): container finished" podID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerID="3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608" exitCode=0 Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.513260 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerDied","Data":"3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608"} Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.513280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltvf2" event={"ID":"4b077eed-9113-4db8-9ba9-8b5d960ad872","Type":"ContainerDied","Data":"da2279b34a21db6da5bd5d05a4788c7648729cfc9905e8a5a8aa9b4710680c98"} Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.513299 4926 scope.go:117] "RemoveContainer" containerID="3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.513454 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltvf2" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.515788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" event={"ID":"37e8795d-6050-4b61-9282-f3af17a926c8","Type":"ContainerStarted","Data":"411f6556c2086550202bb7b8320d0070ea7c0f18b8e05241bba09989f5ba5830"} Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.515967 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.534598 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2vnrc" podStartSLOduration=1.7805802389999998 podStartE2EDuration="7.534574814s" podCreationTimestamp="2025-10-07 21:09:50 +0000 UTC" firstStartedPulling="2025-10-07 21:09:51.450402035 +0000 UTC m=+841.488983185" lastFinishedPulling="2025-10-07 21:09:57.20439661 +0000 UTC m=+847.242977760" observedRunningTime="2025-10-07 21:09:57.527796218 +0000 UTC m=+847.566377378" watchObservedRunningTime="2025-10-07 21:09:57.534574814 +0000 UTC m=+847.573155974" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.539563 4926 scope.go:117] "RemoveContainer" containerID="002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.564808 4926 scope.go:117] "RemoveContainer" containerID="cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.585079 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" podStartSLOduration=1.811277649 podStartE2EDuration="7.585057535s" podCreationTimestamp="2025-10-07 21:09:50 +0000 UTC" firstStartedPulling="2025-10-07 21:09:51.502767822 +0000 UTC m=+841.541348972" lastFinishedPulling="2025-10-07 21:09:57.276547708 +0000 UTC m=+847.315128858" observedRunningTime="2025-10-07 21:09:57.579572626 +0000 UTC m=+847.618153816" watchObservedRunningTime="2025-10-07 21:09:57.585057535 +0000 UTC m=+847.623638695" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.600345 4926 scope.go:117] "RemoveContainer" containerID="3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608" Oct 07 21:09:57 crc kubenswrapper[4926]: E1007 21:09:57.600908 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608\": container with ID starting with 3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608 not found: ID does not exist" containerID="3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.600958 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608"} err="failed to get container status \"3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608\": rpc error: code = NotFound desc = could not find container \"3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608\": container with ID starting with 3c9db34c03527418378ad59eef87bec8dc6911f1da330193f926442df04df608 not found: ID does not exist" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.600990 4926 scope.go:117] "RemoveContainer" containerID="002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.601050 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ltvf2"] Oct 07 21:09:57 crc kubenswrapper[4926]: E1007 21:09:57.601507 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0\": container with ID starting with 002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0 not found: ID does not exist" containerID="002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.601539 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0"} err="failed to get container status \"002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0\": rpc error: code = NotFound desc = could not find container \"002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0\": container with ID starting with 002d82800f163ec6b86f2b652050bfcc4f083a1a2e2a370ffa6ae29d1a54bac0 not found: ID does not exist" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.601559 4926 scope.go:117] "RemoveContainer" containerID="cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba" Oct 07 21:09:57 crc kubenswrapper[4926]: E1007 21:09:57.602723 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba\": container with ID starting with cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba not found: ID does not exist" containerID="cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.602758 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba"} err="failed to get container status \"cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba\": rpc error: code = NotFound desc = could not find container \"cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba\": container with ID starting with cf19995939cd27e862887849def7c2464624d15d614f72b431fb384966ee1cba not found: ID does not exist" Oct 07 21:09:57 crc kubenswrapper[4926]: I1007 21:09:57.606504 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ltvf2"] Oct 07 21:09:58 crc kubenswrapper[4926]: I1007 21:09:58.690848 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" path="/var/lib/kubelet/pods/4b077eed-9113-4db8-9ba9-8b5d960ad872/volumes" Oct 07 21:10:00 crc kubenswrapper[4926]: I1007 21:10:00.871063 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:10:00 crc kubenswrapper[4926]: I1007 21:10:00.871548 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:10:00 crc kubenswrapper[4926]: I1007 21:10:00.953906 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:10:00 crc kubenswrapper[4926]: I1007 21:10:00.989238 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8578696b-282fd" Oct 07 21:10:02 crc kubenswrapper[4926]: I1007 21:10:02.632435 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:10:04 crc kubenswrapper[4926]: I1007 21:10:04.331905 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2vnrc"] Oct 07 21:10:04 crc kubenswrapper[4926]: I1007 21:10:04.567597 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2vnrc" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="registry-server" containerID="cri-o://ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146" gracePeriod=2 Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.060514 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.136772 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-catalog-content\") pod \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.137292 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh72n\" (UniqueName: \"kubernetes.io/projected/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-kube-api-access-rh72n\") pod \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.137448 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-utilities\") pod \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\" (UID: \"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c\") " Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.138014 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-utilities" (OuterVolumeSpecName: "utilities") pod "d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" (UID: "d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.142825 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-kube-api-access-rh72n" (OuterVolumeSpecName: "kube-api-access-rh72n") pod "d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" (UID: "d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c"). InnerVolumeSpecName "kube-api-access-rh72n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.191417 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" (UID: "d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.238381 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.238418 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.238430 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh72n\" (UniqueName: \"kubernetes.io/projected/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c-kube-api-access-rh72n\") on node \"crc\" DevicePath \"\"" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.579618 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerID="ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146" exitCode=0 Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.579696 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vnrc" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.581168 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vnrc" event={"ID":"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c","Type":"ContainerDied","Data":"ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146"} Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.581445 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vnrc" event={"ID":"d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c","Type":"ContainerDied","Data":"c07ce3cc7ccdbb41fb5fa66fa29cbd2333fe7ebfa9e21a5c6413d5c83bc3f241"} Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.581563 4926 scope.go:117] "RemoveContainer" containerID="ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.619827 4926 scope.go:117] "RemoveContainer" containerID="60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.637562 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2vnrc"] Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.644351 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2vnrc"] Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.658630 4926 scope.go:117] "RemoveContainer" containerID="742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.678300 4926 scope.go:117] "RemoveContainer" containerID="ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146" Oct 07 21:10:05 crc kubenswrapper[4926]: E1007 21:10:05.679170 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146\": container with ID starting with ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146 not found: ID does not exist" containerID="ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.679323 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146"} err="failed to get container status \"ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146\": rpc error: code = NotFound desc = could not find container \"ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146\": container with ID starting with ec4ec1e9795eb249b86e23d35c745cb83f967aaeb125485f4812bdc31e857146 not found: ID does not exist" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.679358 4926 scope.go:117] "RemoveContainer" containerID="60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae" Oct 07 21:10:05 crc kubenswrapper[4926]: E1007 21:10:05.680098 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae\": container with ID starting with 60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae not found: ID does not exist" containerID="60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.680138 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae"} err="failed to get container status \"60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae\": rpc error: code = NotFound desc = could not find container \"60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae\": container with ID starting with 60e677cce909d142c2fb17ddc569c6061ef89af97611b232254ea6f77d7ac6ae not found: ID does not exist" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.680165 4926 scope.go:117] "RemoveContainer" containerID="742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754" Oct 07 21:10:05 crc kubenswrapper[4926]: E1007 21:10:05.684436 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754\": container with ID starting with 742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754 not found: ID does not exist" containerID="742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754" Oct 07 21:10:05 crc kubenswrapper[4926]: I1007 21:10:05.684503 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754"} err="failed to get container status \"742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754\": rpc error: code = NotFound desc = could not find container \"742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754\": container with ID starting with 742e7dab178f92435e1173306ff07f231c9cea75599a66c27180e3ae201b3754 not found: ID does not exist" Oct 07 21:10:06 crc kubenswrapper[4926]: I1007 21:10:06.687299 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" path="/var/lib/kubelet/pods/d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c/volumes" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.804590 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns"] Oct 07 21:10:17 crc kubenswrapper[4926]: E1007 21:10:17.805378 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="registry-server" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805392 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="registry-server" Oct 07 21:10:17 crc kubenswrapper[4926]: E1007 21:10:17.805403 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="extract-utilities" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805411 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="extract-utilities" Oct 07 21:10:17 crc kubenswrapper[4926]: E1007 21:10:17.805423 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="extract-utilities" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805432 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="extract-utilities" Oct 07 21:10:17 crc kubenswrapper[4926]: E1007 21:10:17.805448 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="extract-content" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805456 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="extract-content" Oct 07 21:10:17 crc kubenswrapper[4926]: E1007 21:10:17.805467 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="extract-content" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805475 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="extract-content" Oct 07 21:10:17 crc kubenswrapper[4926]: E1007 21:10:17.805486 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="registry-server" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805495 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="registry-server" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805628 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5fe8d32-1ba9-491c-b00f-36ca1a3bbf1c" containerName="registry-server" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.805648 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b077eed-9113-4db8-9ba9-8b5d960ad872" containerName="registry-server" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.806381 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.810091 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-8smlt" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.821464 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.823164 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.828120 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.831082 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-l6vpz" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.847607 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.848920 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.851508 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-rlwf6" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.860997 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.877643 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.878554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.882499 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-p7nnz" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.888944 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.889831 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.895000 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-w8wnn" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.895237 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.914604 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.921054 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.930045 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.931074 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.931778 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99xp9\" (UniqueName: \"kubernetes.io/projected/313c922a-ffe2-4d92-a221-bde117d0f196-kube-api-access-99xp9\") pod \"cinder-operator-controller-manager-7d4d4f8d-bljj9\" (UID: \"313c922a-ffe2-4d92-a221-bde117d0f196\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.931852 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65dd7\" (UniqueName: \"kubernetes.io/projected/fea578e9-bbed-446d-beb9-5fcab1884b60-kube-api-access-65dd7\") pod \"barbican-operator-controller-manager-58c4cd55f4-7ltns\" (UID: \"fea578e9-bbed-446d-beb9-5fcab1884b60\") " pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.931874 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbj2q\" (UniqueName: \"kubernetes.io/projected/a2ab6887-9ac0-4beb-8da0-8cb27da2ff66-kube-api-access-kbj2q\") pod \"designate-operator-controller-manager-75dfd9b554-pwsh8\" (UID: \"a2ab6887-9ac0-4beb-8da0-8cb27da2ff66\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.933450 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-tzbfw" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.957418 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf"] Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.958367 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.960590 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.960806 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-wcrcm" Oct 07 21:10:17 crc kubenswrapper[4926]: I1007 21:10:17.984051 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.002394 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.003545 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.009175 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ggx9v" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.013050 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.018274 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.019327 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.027744 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-dqqs4" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.032884 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbcgt\" (UniqueName: \"kubernetes.io/projected/bd230ff4-3566-43dc-be73-46cfe038bd76-kube-api-access-gbcgt\") pod \"glance-operator-controller-manager-5dc44df7d5-jnbmb\" (UID: \"bd230ff4-3566-43dc-be73-46cfe038bd76\") " pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.032952 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqxpw\" (UniqueName: \"kubernetes.io/projected/6cb125b4-a707-4506-80fc-76bc981f1560-kube-api-access-mqxpw\") pod \"horizon-operator-controller-manager-76d5b87f47-nm2mk\" (UID: \"6cb125b4-a707-4506-80fc-76bc981f1560\") " pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.032977 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv959\" (UniqueName: \"kubernetes.io/projected/5517a3d5-4f45-44db-9552-fcd8de7356bf-kube-api-access-jv959\") pod \"infra-operator-controller-manager-658588b8c9-cgkzf\" (UID: \"5517a3d5-4f45-44db-9552-fcd8de7356bf\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.033007 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99xp9\" (UniqueName: \"kubernetes.io/projected/313c922a-ffe2-4d92-a221-bde117d0f196-kube-api-access-99xp9\") pod \"cinder-operator-controller-manager-7d4d4f8d-bljj9\" (UID: \"313c922a-ffe2-4d92-a221-bde117d0f196\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.033059 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65dd7\" (UniqueName: \"kubernetes.io/projected/fea578e9-bbed-446d-beb9-5fcab1884b60-kube-api-access-65dd7\") pod \"barbican-operator-controller-manager-58c4cd55f4-7ltns\" (UID: \"fea578e9-bbed-446d-beb9-5fcab1884b60\") " pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.033075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbj2q\" (UniqueName: \"kubernetes.io/projected/a2ab6887-9ac0-4beb-8da0-8cb27da2ff66-kube-api-access-kbj2q\") pod \"designate-operator-controller-manager-75dfd9b554-pwsh8\" (UID: \"a2ab6887-9ac0-4beb-8da0-8cb27da2ff66\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.033097 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jh6x\" (UniqueName: \"kubernetes.io/projected/fc46d600-5c99-4ce4-9beb-06376dbacd8a-kube-api-access-5jh6x\") pod \"heat-operator-controller-manager-54b4974c45-mwlfv\" (UID: \"fc46d600-5c99-4ce4-9beb-06376dbacd8a\") " pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.033120 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5517a3d5-4f45-44db-9552-fcd8de7356bf-cert\") pod \"infra-operator-controller-manager-658588b8c9-cgkzf\" (UID: \"5517a3d5-4f45-44db-9552-fcd8de7356bf\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.045240 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.046184 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.049460 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.067086 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-xfxpf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.084144 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99xp9\" (UniqueName: \"kubernetes.io/projected/313c922a-ffe2-4d92-a221-bde117d0f196-kube-api-access-99xp9\") pod \"cinder-operator-controller-manager-7d4d4f8d-bljj9\" (UID: \"313c922a-ffe2-4d92-a221-bde117d0f196\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.095293 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.100576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbj2q\" (UniqueName: \"kubernetes.io/projected/a2ab6887-9ac0-4beb-8da0-8cb27da2ff66-kube-api-access-kbj2q\") pod \"designate-operator-controller-manager-75dfd9b554-pwsh8\" (UID: \"a2ab6887-9ac0-4beb-8da0-8cb27da2ff66\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.138961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jh6x\" (UniqueName: \"kubernetes.io/projected/fc46d600-5c99-4ce4-9beb-06376dbacd8a-kube-api-access-5jh6x\") pod \"heat-operator-controller-manager-54b4974c45-mwlfv\" (UID: \"fc46d600-5c99-4ce4-9beb-06376dbacd8a\") " pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139003 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5517a3d5-4f45-44db-9552-fcd8de7356bf-cert\") pod \"infra-operator-controller-manager-658588b8c9-cgkzf\" (UID: \"5517a3d5-4f45-44db-9552-fcd8de7356bf\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139410 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4km4z\" (UniqueName: \"kubernetes.io/projected/cefed49a-2a06-49a6-917e-ff05d6457bca-kube-api-access-4km4z\") pod \"keystone-operator-controller-manager-7b5ccf6d9c-wxrzg\" (UID: \"cefed49a-2a06-49a6-917e-ff05d6457bca\") " pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139439 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbcgt\" (UniqueName: \"kubernetes.io/projected/bd230ff4-3566-43dc-be73-46cfe038bd76-kube-api-access-gbcgt\") pod \"glance-operator-controller-manager-5dc44df7d5-jnbmb\" (UID: \"bd230ff4-3566-43dc-be73-46cfe038bd76\") " pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139471 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6pjx\" (UniqueName: \"kubernetes.io/projected/d1f7536d-1b34-4d39-9249-b9959aa2a00f-kube-api-access-c6pjx\") pod \"ironic-operator-controller-manager-649675d675-zz2jk\" (UID: \"d1f7536d-1b34-4d39-9249-b9959aa2a00f\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139516 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqxpw\" (UniqueName: \"kubernetes.io/projected/6cb125b4-a707-4506-80fc-76bc981f1560-kube-api-access-mqxpw\") pod \"horizon-operator-controller-manager-76d5b87f47-nm2mk\" (UID: \"6cb125b4-a707-4506-80fc-76bc981f1560\") " pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139542 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv959\" (UniqueName: \"kubernetes.io/projected/5517a3d5-4f45-44db-9552-fcd8de7356bf-kube-api-access-jv959\") pod \"infra-operator-controller-manager-658588b8c9-cgkzf\" (UID: \"5517a3d5-4f45-44db-9552-fcd8de7356bf\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.139581 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnqr2\" (UniqueName: \"kubernetes.io/projected/bd7d6078-1c87-459e-b83e-c408537f82a1-kube-api-access-wnqr2\") pod \"manila-operator-controller-manager-65d89cfd9f-tmh94\" (UID: \"bd7d6078-1c87-459e-b83e-c408537f82a1\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.141862 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65dd7\" (UniqueName: \"kubernetes.io/projected/fea578e9-bbed-446d-beb9-5fcab1884b60-kube-api-access-65dd7\") pod \"barbican-operator-controller-manager-58c4cd55f4-7ltns\" (UID: \"fea578e9-bbed-446d-beb9-5fcab1884b60\") " pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.149576 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.167776 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.168721 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.170300 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5517a3d5-4f45-44db-9552-fcd8de7356bf-cert\") pod \"infra-operator-controller-manager-658588b8c9-cgkzf\" (UID: \"5517a3d5-4f45-44db-9552-fcd8de7356bf\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.193629 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.195341 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbcgt\" (UniqueName: \"kubernetes.io/projected/bd230ff4-3566-43dc-be73-46cfe038bd76-kube-api-access-gbcgt\") pod \"glance-operator-controller-manager-5dc44df7d5-jnbmb\" (UID: \"bd230ff4-3566-43dc-be73-46cfe038bd76\") " pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.199066 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqxpw\" (UniqueName: \"kubernetes.io/projected/6cb125b4-a707-4506-80fc-76bc981f1560-kube-api-access-mqxpw\") pod \"horizon-operator-controller-manager-76d5b87f47-nm2mk\" (UID: \"6cb125b4-a707-4506-80fc-76bc981f1560\") " pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.199802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jh6x\" (UniqueName: \"kubernetes.io/projected/fc46d600-5c99-4ce4-9beb-06376dbacd8a-kube-api-access-5jh6x\") pod \"heat-operator-controller-manager-54b4974c45-mwlfv\" (UID: \"fc46d600-5c99-4ce4-9beb-06376dbacd8a\") " pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.201980 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv959\" (UniqueName: \"kubernetes.io/projected/5517a3d5-4f45-44db-9552-fcd8de7356bf-kube-api-access-jv959\") pod \"infra-operator-controller-manager-658588b8c9-cgkzf\" (UID: \"5517a3d5-4f45-44db-9552-fcd8de7356bf\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.214241 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.215390 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.220700 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tq49n" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.223012 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.224662 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.235107 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.240509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4km4z\" (UniqueName: \"kubernetes.io/projected/cefed49a-2a06-49a6-917e-ff05d6457bca-kube-api-access-4km4z\") pod \"keystone-operator-controller-manager-7b5ccf6d9c-wxrzg\" (UID: \"cefed49a-2a06-49a6-917e-ff05d6457bca\") " pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.240553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6pjx\" (UniqueName: \"kubernetes.io/projected/d1f7536d-1b34-4d39-9249-b9959aa2a00f-kube-api-access-c6pjx\") pod \"ironic-operator-controller-manager-649675d675-zz2jk\" (UID: \"d1f7536d-1b34-4d39-9249-b9959aa2a00f\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.240623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnqr2\" (UniqueName: \"kubernetes.io/projected/bd7d6078-1c87-459e-b83e-c408537f82a1-kube-api-access-wnqr2\") pod \"manila-operator-controller-manager-65d89cfd9f-tmh94\" (UID: \"bd7d6078-1c87-459e-b83e-c408537f82a1\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.252470 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.255254 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.256746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4km4z\" (UniqueName: \"kubernetes.io/projected/cefed49a-2a06-49a6-917e-ff05d6457bca-kube-api-access-4km4z\") pod \"keystone-operator-controller-manager-7b5ccf6d9c-wxrzg\" (UID: \"cefed49a-2a06-49a6-917e-ff05d6457bca\") " pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.257911 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.257995 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6pjx\" (UniqueName: \"kubernetes.io/projected/d1f7536d-1b34-4d39-9249-b9959aa2a00f-kube-api-access-c6pjx\") pod \"ironic-operator-controller-manager-649675d675-zz2jk\" (UID: \"d1f7536d-1b34-4d39-9249-b9959aa2a00f\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.258134 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnqr2\" (UniqueName: \"kubernetes.io/projected/bd7d6078-1c87-459e-b83e-c408537f82a1-kube-api-access-wnqr2\") pod \"manila-operator-controller-manager-65d89cfd9f-tmh94\" (UID: \"bd7d6078-1c87-459e-b83e-c408537f82a1\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.259924 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-xqjmk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.260770 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.267732 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.268996 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.270458 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-t8tz9" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.273568 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.286298 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.295422 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.296957 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.299783 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qw6x4" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.303554 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.315800 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.316952 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.318994 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9xq9g" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.319293 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.323554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.330284 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.339392 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.339509 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.340401 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.341875 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbsmx\" (UniqueName: \"kubernetes.io/projected/0d0347d8-c6bc-4bd5-b75a-28ba94906a6d-kube-api-access-hbsmx\") pod \"nova-operator-controller-manager-7c7fc454ff-4xz96\" (UID: \"0d0347d8-c6bc-4bd5-b75a-28ba94906a6d\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.341942 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vwmv\" (UniqueName: \"kubernetes.io/projected/2d16f135-1997-4bb6-9cb3-dae655640a01-kube-api-access-5vwmv\") pod \"octavia-operator-controller-manager-7468f855d8-rxdc8\" (UID: \"2d16f135-1997-4bb6-9cb3-dae655640a01\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.341976 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g2j8\" (UniqueName: \"kubernetes.io/projected/ce47a525-41f3-4beb-b9b3-28d85831fdda-kube-api-access-4g2j8\") pod \"neutron-operator-controller-manager-8d984cc4d-tjfcp\" (UID: \"ce47a525-41f3-4beb-b9b3-28d85831fdda\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.342280 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-4k4rj" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.345563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7588q\" (UniqueName: \"kubernetes.io/projected/09117b6f-d508-460f-8848-ad0ba4f7c1be-kube-api-access-7588q\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-7spjf\" (UID: \"09117b6f-d508-460f-8848-ad0ba4f7c1be\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.360534 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.371830 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.375015 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.385169 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w9bk4" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.398103 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.401119 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.406522 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-d9zzf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.428667 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.437498 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.443316 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.446848 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.448146 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.450718 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t22sx\" (UniqueName: \"kubernetes.io/projected/35b344ad-2ef7-48e3-8643-e93e974fb326-kube-api-access-t22sx\") pod \"swift-operator-controller-manager-6859f9b676-59xh2\" (UID: \"35b344ad-2ef7-48e3-8643-e93e974fb326\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.450873 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11159753-a963-44c9-bb63-53340a200de7-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.450940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7588q\" (UniqueName: \"kubernetes.io/projected/09117b6f-d508-460f-8848-ad0ba4f7c1be-kube-api-access-7588q\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-7spjf\" (UID: \"09117b6f-d508-460f-8848-ad0ba4f7c1be\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.450996 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwdnq\" (UniqueName: \"kubernetes.io/projected/11159753-a963-44c9-bb63-53340a200de7-kube-api-access-fwdnq\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.451042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cgcl\" (UniqueName: \"kubernetes.io/projected/d039dde6-16e1-42dd-a19f-b564a3939837-kube-api-access-7cgcl\") pod \"placement-operator-controller-manager-54689d9f88-47m5t\" (UID: \"d039dde6-16e1-42dd-a19f-b564a3939837\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.451068 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbsmx\" (UniqueName: \"kubernetes.io/projected/0d0347d8-c6bc-4bd5-b75a-28ba94906a6d-kube-api-access-hbsmx\") pod \"nova-operator-controller-manager-7c7fc454ff-4xz96\" (UID: \"0d0347d8-c6bc-4bd5-b75a-28ba94906a6d\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.451161 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vwmv\" (UniqueName: \"kubernetes.io/projected/2d16f135-1997-4bb6-9cb3-dae655640a01-kube-api-access-5vwmv\") pod \"octavia-operator-controller-manager-7468f855d8-rxdc8\" (UID: \"2d16f135-1997-4bb6-9cb3-dae655640a01\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.451229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g2j8\" (UniqueName: \"kubernetes.io/projected/ce47a525-41f3-4beb-b9b3-28d85831fdda-kube-api-access-4g2j8\") pod \"neutron-operator-controller-manager-8d984cc4d-tjfcp\" (UID: \"ce47a525-41f3-4beb-b9b3-28d85831fdda\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.451306 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wj7\" (UniqueName: \"kubernetes.io/projected/f490b56a-2c4e-49ea-93fc-f2d8fb9454fd-kube-api-access-j9wj7\") pod \"ovn-operator-controller-manager-6d8b6f9b9-tsklh\" (UID: \"f490b56a-2c4e-49ea-93fc-f2d8fb9454fd\") " pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.454877 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8zkh9" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.476580 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7588q\" (UniqueName: \"kubernetes.io/projected/09117b6f-d508-460f-8848-ad0ba4f7c1be-kube-api-access-7588q\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-7spjf\" (UID: \"09117b6f-d508-460f-8848-ad0ba4f7c1be\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.479578 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.482216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vwmv\" (UniqueName: \"kubernetes.io/projected/2d16f135-1997-4bb6-9cb3-dae655640a01-kube-api-access-5vwmv\") pod \"octavia-operator-controller-manager-7468f855d8-rxdc8\" (UID: \"2d16f135-1997-4bb6-9cb3-dae655640a01\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.483580 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbsmx\" (UniqueName: \"kubernetes.io/projected/0d0347d8-c6bc-4bd5-b75a-28ba94906a6d-kube-api-access-hbsmx\") pod \"nova-operator-controller-manager-7c7fc454ff-4xz96\" (UID: \"0d0347d8-c6bc-4bd5-b75a-28ba94906a6d\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.488253 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g2j8\" (UniqueName: \"kubernetes.io/projected/ce47a525-41f3-4beb-b9b3-28d85831fdda-kube-api-access-4g2j8\") pod \"neutron-operator-controller-manager-8d984cc4d-tjfcp\" (UID: \"ce47a525-41f3-4beb-b9b3-28d85831fdda\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.501722 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.507805 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.507907 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.513433 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-zl9wc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.552686 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.553967 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11159753-a963-44c9-bb63-53340a200de7-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.554029 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwdnq\" (UniqueName: \"kubernetes.io/projected/11159753-a963-44c9-bb63-53340a200de7-kube-api-access-fwdnq\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.554053 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cgcl\" (UniqueName: \"kubernetes.io/projected/d039dde6-16e1-42dd-a19f-b564a3939837-kube-api-access-7cgcl\") pod \"placement-operator-controller-manager-54689d9f88-47m5t\" (UID: \"d039dde6-16e1-42dd-a19f-b564a3939837\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.554094 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl4dq\" (UniqueName: \"kubernetes.io/projected/659a6e11-31f5-4a95-9eb1-3f61d8891ace-kube-api-access-rl4dq\") pod \"telemetry-operator-controller-manager-5d4d74dd89-w2bgf\" (UID: \"659a6e11-31f5-4a95-9eb1-3f61d8891ace\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.554138 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wj7\" (UniqueName: \"kubernetes.io/projected/f490b56a-2c4e-49ea-93fc-f2d8fb9454fd-kube-api-access-j9wj7\") pod \"ovn-operator-controller-manager-6d8b6f9b9-tsklh\" (UID: \"f490b56a-2c4e-49ea-93fc-f2d8fb9454fd\") " pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.554159 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t22sx\" (UniqueName: \"kubernetes.io/projected/35b344ad-2ef7-48e3-8643-e93e974fb326-kube-api-access-t22sx\") pod \"swift-operator-controller-manager-6859f9b676-59xh2\" (UID: \"35b344ad-2ef7-48e3-8643-e93e974fb326\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:18 crc kubenswrapper[4926]: E1007 21:10:18.554564 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 21:10:18 crc kubenswrapper[4926]: E1007 21:10:18.554610 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11159753-a963-44c9-bb63-53340a200de7-cert podName:11159753-a963-44c9-bb63-53340a200de7 nodeName:}" failed. No retries permitted until 2025-10-07 21:10:19.054595484 +0000 UTC m=+869.093176634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/11159753-a963-44c9-bb63-53340a200de7-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" (UID: "11159753-a963-44c9-bb63-53340a200de7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.583136 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.587392 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cgcl\" (UniqueName: \"kubernetes.io/projected/d039dde6-16e1-42dd-a19f-b564a3939837-kube-api-access-7cgcl\") pod \"placement-operator-controller-manager-54689d9f88-47m5t\" (UID: \"d039dde6-16e1-42dd-a19f-b564a3939837\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.602662 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.607862 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwdnq\" (UniqueName: \"kubernetes.io/projected/11159753-a963-44c9-bb63-53340a200de7-kube-api-access-fwdnq\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.609637 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.609688 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wj7\" (UniqueName: \"kubernetes.io/projected/f490b56a-2c4e-49ea-93fc-f2d8fb9454fd-kube-api-access-j9wj7\") pod \"ovn-operator-controller-manager-6d8b6f9b9-tsklh\" (UID: \"f490b56a-2c4e-49ea-93fc-f2d8fb9454fd\") " pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.612512 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.630853 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-fcdj2" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.643238 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t22sx\" (UniqueName: \"kubernetes.io/projected/35b344ad-2ef7-48e3-8643-e93e974fb326-kube-api-access-t22sx\") pod \"swift-operator-controller-manager-6859f9b676-59xh2\" (UID: \"35b344ad-2ef7-48e3-8643-e93e974fb326\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.648788 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.669026 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.670951 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46qsz\" (UniqueName: \"kubernetes.io/projected/f825a6cc-9414-44f5-98a2-df2c3f1e825c-kube-api-access-46qsz\") pod \"test-operator-controller-manager-5cd5cb47d7-rtwpd\" (UID: \"f825a6cc-9414-44f5-98a2-df2c3f1e825c\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.671008 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl4dq\" (UniqueName: \"kubernetes.io/projected/659a6e11-31f5-4a95-9eb1-3f61d8891ace-kube-api-access-rl4dq\") pod \"telemetry-operator-controller-manager-5d4d74dd89-w2bgf\" (UID: \"659a6e11-31f5-4a95-9eb1-3f61d8891ace\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.698674 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.716944 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.720626 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl4dq\" (UniqueName: \"kubernetes.io/projected/659a6e11-31f5-4a95-9eb1-3f61d8891ace-kube-api-access-rl4dq\") pod \"telemetry-operator-controller-manager-5d4d74dd89-w2bgf\" (UID: \"659a6e11-31f5-4a95-9eb1-3f61d8891ace\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.743555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.762734 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.766495 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.766578 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.770320 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-qf7sd" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.770508 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.772492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdzbp\" (UniqueName: \"kubernetes.io/projected/45317f9d-c1aa-4f42-b52f-f910a7082d1c-kube-api-access-cdzbp\") pod \"watcher-operator-controller-manager-77fdd5cc9b-q8mkp\" (UID: \"45317f9d-c1aa-4f42-b52f-f910a7082d1c\") " pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.772551 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46qsz\" (UniqueName: \"kubernetes.io/projected/f825a6cc-9414-44f5-98a2-df2c3f1e825c-kube-api-access-46qsz\") pod \"test-operator-controller-manager-5cd5cb47d7-rtwpd\" (UID: \"f825a6cc-9414-44f5-98a2-df2c3f1e825c\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.775654 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.776829 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.779720 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-mkw97" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.791158 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.797051 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.798353 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.801390 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46qsz\" (UniqueName: \"kubernetes.io/projected/f825a6cc-9414-44f5-98a2-df2c3f1e825c-kube-api-access-46qsz\") pod \"test-operator-controller-manager-5cd5cb47d7-rtwpd\" (UID: \"f825a6cc-9414-44f5-98a2-df2c3f1e825c\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.815562 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.829023 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.880131 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcm2d\" (UniqueName: \"kubernetes.io/projected/fb36bed3-c7f6-43ad-b250-da42097b673e-kube-api-access-tcm2d\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5\" (UID: \"fb36bed3-c7f6-43ad-b250-da42097b673e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.880209 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66vh4\" (UniqueName: \"kubernetes.io/projected/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-kube-api-access-66vh4\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.880248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdzbp\" (UniqueName: \"kubernetes.io/projected/45317f9d-c1aa-4f42-b52f-f910a7082d1c-kube-api-access-cdzbp\") pod \"watcher-operator-controller-manager-77fdd5cc9b-q8mkp\" (UID: \"45317f9d-c1aa-4f42-b52f-f910a7082d1c\") " pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.880285 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-cert\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.902727 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdzbp\" (UniqueName: \"kubernetes.io/projected/45317f9d-c1aa-4f42-b52f-f910a7082d1c-kube-api-access-cdzbp\") pod \"watcher-operator-controller-manager-77fdd5cc9b-q8mkp\" (UID: \"45317f9d-c1aa-4f42-b52f-f910a7082d1c\") " pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.947761 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9"] Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.982983 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcm2d\" (UniqueName: \"kubernetes.io/projected/fb36bed3-c7f6-43ad-b250-da42097b673e-kube-api-access-tcm2d\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5\" (UID: \"fb36bed3-c7f6-43ad-b250-da42097b673e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.983421 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66vh4\" (UniqueName: \"kubernetes.io/projected/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-kube-api-access-66vh4\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.983454 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-cert\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:18 crc kubenswrapper[4926]: E1007 21:10:18.983562 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 07 21:10:18 crc kubenswrapper[4926]: E1007 21:10:18.983608 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-cert podName:78f2ab6d-0297-49e4-91ff-ae7c8251fb0e nodeName:}" failed. No retries permitted until 2025-10-07 21:10:19.483592738 +0000 UTC m=+869.522173888 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-cert") pod "openstack-operator-controller-manager-9554d4fbd-2cvhj" (UID: "78f2ab6d-0297-49e4-91ff-ae7c8251fb0e") : secret "webhook-server-cert" not found Oct 07 21:10:18 crc kubenswrapper[4926]: I1007 21:10:18.986635 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.009214 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66vh4\" (UniqueName: \"kubernetes.io/projected/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-kube-api-access-66vh4\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.009477 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcm2d\" (UniqueName: \"kubernetes.io/projected/fb36bed3-c7f6-43ad-b250-da42097b673e-kube-api-access-tcm2d\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5\" (UID: \"fb36bed3-c7f6-43ad-b250-da42097b673e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.016743 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.027663 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd230ff4_3566_43dc_be73_46cfe038bd76.slice/crio-ae4dbfc54810f85732643674af6c3480609b2eec7fad0ba6b27d04daa8681afd WatchSource:0}: Error finding container ae4dbfc54810f85732643674af6c3480609b2eec7fad0ba6b27d04daa8681afd: Status 404 returned error can't find the container with id ae4dbfc54810f85732643674af6c3480609b2eec7fad0ba6b27d04daa8681afd Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.086568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11159753-a963-44c9-bb63-53340a200de7-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.101082 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11159753-a963-44c9-bb63-53340a200de7-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc\" (UID: \"11159753-a963-44c9-bb63-53340a200de7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.122427 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.184679 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.188434 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.197398 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.212618 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.273097 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.309724 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd7d6078_1c87_459e_b83e_c408537f82a1.slice/crio-a6e344f4c2dedad6232310943a3ba3f04e8d915d82ebfec9cf7af6099073322c WatchSource:0}: Error finding container a6e344f4c2dedad6232310943a3ba3f04e8d915d82ebfec9cf7af6099073322c: Status 404 returned error can't find the container with id a6e344f4c2dedad6232310943a3ba3f04e8d915d82ebfec9cf7af6099073322c Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.338142 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1f7536d_1b34_4d39_9249_b9959aa2a00f.slice/crio-3fbf7868ed8018be20400c3d3e7f45e643a7bb5e4f92100121a1afad3e938336 WatchSource:0}: Error finding container 3fbf7868ed8018be20400c3d3e7f45e643a7bb5e4f92100121a1afad3e938336: Status 404 returned error can't find the container with id 3fbf7868ed8018be20400c3d3e7f45e643a7bb5e4f92100121a1afad3e938336 Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.415675 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.471393 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.476386 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ab6887_9ac0_4beb_8da0_8cb27da2ff66.slice/crio-01277d4dbfb8b9d095186ea172eea2cd88ca8effa9c5cc80d8ddb3aee4828da6 WatchSource:0}: Error finding container 01277d4dbfb8b9d095186ea172eea2cd88ca8effa9c5cc80d8ddb3aee4828da6: Status 404 returned error can't find the container with id 01277d4dbfb8b9d095186ea172eea2cd88ca8effa9c5cc80d8ddb3aee4828da6 Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.493554 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-cert\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.499054 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78f2ab6d-0297-49e4-91ff-ae7c8251fb0e-cert\") pod \"openstack-operator-controller-manager-9554d4fbd-2cvhj\" (UID: \"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e\") " pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.703060 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.730901 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" event={"ID":"313c922a-ffe2-4d92-a221-bde117d0f196","Type":"ContainerStarted","Data":"447a625c8b04f4c471a741f1c96bbb74b0c0a873714c3e654651ad9c15317cd0"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.744810 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" event={"ID":"a2ab6887-9ac0-4beb-8da0-8cb27da2ff66","Type":"ContainerStarted","Data":"01277d4dbfb8b9d095186ea172eea2cd88ca8effa9c5cc80d8ddb3aee4828da6"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.750777 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" event={"ID":"bd230ff4-3566-43dc-be73-46cfe038bd76","Type":"ContainerStarted","Data":"ae4dbfc54810f85732643674af6c3480609b2eec7fad0ba6b27d04daa8681afd"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.753582 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" event={"ID":"bd7d6078-1c87-459e-b83e-c408537f82a1","Type":"ContainerStarted","Data":"a6e344f4c2dedad6232310943a3ba3f04e8d915d82ebfec9cf7af6099073322c"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.759335 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" event={"ID":"fea578e9-bbed-446d-beb9-5fcab1884b60","Type":"ContainerStarted","Data":"cb068749a2e42bbf77f84bfc54e55f3125d7f3d465f7a509ce4482777065c54e"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.762936 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" event={"ID":"fc46d600-5c99-4ce4-9beb-06376dbacd8a","Type":"ContainerStarted","Data":"a731f615e0421fb47b680c1a8a61233850f3b568df054d517e7d54bab50f48e9"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.764279 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" event={"ID":"d1f7536d-1b34-4d39-9249-b9959aa2a00f","Type":"ContainerStarted","Data":"3fbf7868ed8018be20400c3d3e7f45e643a7bb5e4f92100121a1afad3e938336"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.766532 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" event={"ID":"5517a3d5-4f45-44db-9552-fcd8de7356bf","Type":"ContainerStarted","Data":"e766042b00e1f85da417ca07de3de3a119340cf126c650bd204e771522ec1b29"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.767781 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" event={"ID":"6cb125b4-a707-4506-80fc-76bc981f1560","Type":"ContainerStarted","Data":"14653157e28eda8ff8e7b60d3487638de84b332c9bc0bd6fc543390fd1711760"} Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.904059 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.910592 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.927747 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d0347d8_c6bc_4bd5_b75a_28ba94906a6d.slice/crio-4243579fe681bf7e2da311571473d687609e181b48f9792e1b6a8ad471228f82 WatchSource:0}: Error finding container 4243579fe681bf7e2da311571473d687609e181b48f9792e1b6a8ad471228f82: Status 404 returned error can't find the container with id 4243579fe681bf7e2da311571473d687609e181b48f9792e1b6a8ad471228f82 Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.930117 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.945446 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d16f135_1997_4bb6_9cb3_dae655640a01.slice/crio-4a79cc19fca0c64299db2dce9d24dad165dd90b8eaf9b891a06f653944ce8958 WatchSource:0}: Error finding container 4a79cc19fca0c64299db2dce9d24dad165dd90b8eaf9b891a06f653944ce8958: Status 404 returned error can't find the container with id 4a79cc19fca0c64299db2dce9d24dad165dd90b8eaf9b891a06f653944ce8958 Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.946695 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp"] Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.952678 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.954057 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce47a525_41f3_4beb_b9b3_28d85831fdda.slice/crio-e768e4b3073e3cf9e170c0b52b1ebb2bade488b58d44488bc5740cf2068910c2 WatchSource:0}: Error finding container e768e4b3073e3cf9e170c0b52b1ebb2bade488b58d44488bc5740cf2068910c2: Status 404 returned error can't find the container with id e768e4b3073e3cf9e170c0b52b1ebb2bade488b58d44488bc5740cf2068910c2 Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.957642 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.960050 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09117b6f_d508_460f_8848_ad0ba4f7c1be.slice/crio-d9c79b887460e0d481c0975d10f96ad3df548046eade0336c3e7b3bc7506f013 WatchSource:0}: Error finding container d9c79b887460e0d481c0975d10f96ad3df548046eade0336c3e7b3bc7506f013: Status 404 returned error can't find the container with id d9c79b887460e0d481c0975d10f96ad3df548046eade0336c3e7b3bc7506f013 Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.962688 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcefed49a_2a06_49a6_917e_ff05d6457bca.slice/crio-2e6442d213d4a7bf090e23a8c08c7a7fb56d092c22f2bd5375c6855b5150eb2c WatchSource:0}: Error finding container 2e6442d213d4a7bf090e23a8c08c7a7fb56d092c22f2bd5375c6855b5150eb2c: Status 404 returned error can't find the container with id 2e6442d213d4a7bf090e23a8c08c7a7fb56d092c22f2bd5375c6855b5150eb2c Oct 07 21:10:19 crc kubenswrapper[4926]: I1007 21:10:19.964328 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2"] Oct 07 21:10:19 crc kubenswrapper[4926]: W1007 21:10:19.973270 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35b344ad_2ef7_48e3_8643_e93e974fb326.slice/crio-02079871637e0e454519b4a31fc5dddc6da7a263a1ca0d9ea4a1093fd4f20d4c WatchSource:0}: Error finding container 02079871637e0e454519b4a31fc5dddc6da7a263a1ca0d9ea4a1093fd4f20d4c: Status 404 returned error can't find the container with id 02079871637e0e454519b4a31fc5dddc6da7a263a1ca0d9ea4a1093fd4f20d4c Oct 07 21:10:19 crc kubenswrapper[4926]: E1007 21:10:19.975552 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t22sx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-59xh2_openstack-operators(35b344ad-2ef7-48e3-8643-e93e974fb326): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.007715 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf"] Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.017791 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd"] Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.030108 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh"] Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.034742 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf490b56a_2c4e_49ea_93fc_f2d8fb9454fd.slice/crio-83d9ab66ab7bdd0fc925eb4a22d0044d99b6a089d5b084f2c0936f38f09e92e7 WatchSource:0}: Error finding container 83d9ab66ab7bdd0fc925eb4a22d0044d99b6a089d5b084f2c0936f38f09e92e7: Status 404 returned error can't find the container with id 83d9ab66ab7bdd0fc925eb4a22d0044d99b6a089d5b084f2c0936f38f09e92e7 Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.040092 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf825a6cc_9414_44f5_98a2_df2c3f1e825c.slice/crio-c42027b15bb25d83cad45404091e72009bec967dad1a844b2d8e953cb191faa7 WatchSource:0}: Error finding container c42027b15bb25d83cad45404091e72009bec967dad1a844b2d8e953cb191faa7: Status 404 returned error can't find the container with id c42027b15bb25d83cad45404091e72009bec967dad1a844b2d8e953cb191faa7 Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.045246 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659a6e11_31f5_4a95_9eb1_3f61d8891ace.slice/crio-6a1fc546fd838cd78247e0d09aa61514319e92f549379425ab226e2584dbc32e WatchSource:0}: Error finding container 6a1fc546fd838cd78247e0d09aa61514319e92f549379425ab226e2584dbc32e: Status 404 returned error can't find the container with id 6a1fc546fd838cd78247e0d09aa61514319e92f549379425ab226e2584dbc32e Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.046885 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46qsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-rtwpd_openstack-operators(f825a6cc-9414-44f5-98a2-df2c3f1e825c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.050993 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rl4dq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d4d74dd89-w2bgf_openstack-operators(659a6e11-31f5-4a95-9eb1-3f61d8891ace): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.261770 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp"] Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.273730 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45317f9d_c1aa_4f42_b52f_f910a7082d1c.slice/crio-d683481ee3d07d70763c6e07a701b948fa65ebb2899aa2a2f2d65cd3d60102ba WatchSource:0}: Error finding container d683481ee3d07d70763c6e07a701b948fa65ebb2899aa2a2f2d65cd3d60102ba: Status 404 returned error can't find the container with id d683481ee3d07d70763c6e07a701b948fa65ebb2899aa2a2f2d65cd3d60102ba Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.278023 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5"] Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.293368 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc"] Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.300659 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb36bed3_c7f6_43ad_b250_da42097b673e.slice/crio-a6648e38345b3b9db82a910696435585cf25cf69ddee2ffbeba33c18174d0121 WatchSource:0}: Error finding container a6648e38345b3b9db82a910696435585cf25cf69ddee2ffbeba33c18174d0121: Status 404 returned error can't find the container with id a6648e38345b3b9db82a910696435585cf25cf69ddee2ffbeba33c18174d0121 Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.308323 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11159753_a963_44c9_bb63_53340a200de7.slice/crio-0025e0cddabfa138d04e1ce88c0684fc2aa49d5e013f2130e6de9783d084ee5e WatchSource:0}: Error finding container 0025e0cddabfa138d04e1ce88c0684fc2aa49d5e013f2130e6de9783d084ee5e: Status 404 returned error can't find the container with id 0025e0cddabfa138d04e1ce88c0684fc2aa49d5e013f2130e6de9783d084ee5e Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.312526 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" podUID="35b344ad-2ef7-48e3-8643-e93e974fb326" Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.312955 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fwdnq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc_openstack-operators(11159753-a963-44c9-bb63-53340a200de7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.313670 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tcm2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5_openstack-operators(fb36bed3-c7f6-43ad-b250-da42097b673e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.314756 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" podUID="fb36bed3-c7f6-43ad-b250-da42097b673e" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.330340 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj"] Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.347083 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" podUID="659a6e11-31f5-4a95-9eb1-3f61d8891ace" Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.356241 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" podUID="f825a6cc-9414-44f5-98a2-df2c3f1e825c" Oct 07 21:10:20 crc kubenswrapper[4926]: W1007 21:10:20.358916 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78f2ab6d_0297_49e4_91ff_ae7c8251fb0e.slice/crio-930e0e3a341080f5b80b1bc32e952a78c1680cf1a86f669e842b8567209af49d WatchSource:0}: Error finding container 930e0e3a341080f5b80b1bc32e952a78c1680cf1a86f669e842b8567209af49d: Status 404 returned error can't find the container with id 930e0e3a341080f5b80b1bc32e952a78c1680cf1a86f669e842b8567209af49d Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.511322 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" podUID="11159753-a963-44c9-bb63-53340a200de7" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.785469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" event={"ID":"ce47a525-41f3-4beb-b9b3-28d85831fdda","Type":"ContainerStarted","Data":"e768e4b3073e3cf9e170c0b52b1ebb2bade488b58d44488bc5740cf2068910c2"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.789545 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" event={"ID":"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e","Type":"ContainerStarted","Data":"db50941f94fe4a2199187f35831f16b35ef119082732a3a01dc759ef39be1418"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.789592 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" event={"ID":"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e","Type":"ContainerStarted","Data":"930e0e3a341080f5b80b1bc32e952a78c1680cf1a86f669e842b8567209af49d"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.793083 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" event={"ID":"f825a6cc-9414-44f5-98a2-df2c3f1e825c","Type":"ContainerStarted","Data":"0b40520799909216a38ea2b6d3e22d9982a4ca0996632ffa731c7001a3382464"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.793114 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" event={"ID":"f825a6cc-9414-44f5-98a2-df2c3f1e825c","Type":"ContainerStarted","Data":"c42027b15bb25d83cad45404091e72009bec967dad1a844b2d8e953cb191faa7"} Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.800605 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" podUID="f825a6cc-9414-44f5-98a2-df2c3f1e825c" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.803319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" event={"ID":"fb36bed3-c7f6-43ad-b250-da42097b673e","Type":"ContainerStarted","Data":"a6648e38345b3b9db82a910696435585cf25cf69ddee2ffbeba33c18174d0121"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.810769 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" event={"ID":"f490b56a-2c4e-49ea-93fc-f2d8fb9454fd","Type":"ContainerStarted","Data":"83d9ab66ab7bdd0fc925eb4a22d0044d99b6a089d5b084f2c0936f38f09e92e7"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.810814 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" event={"ID":"659a6e11-31f5-4a95-9eb1-3f61d8891ace","Type":"ContainerStarted","Data":"eeae74bcf4cde5decd27957fed616fe88a5713438b1ae430cb03fc46ee571433"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.810832 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" event={"ID":"659a6e11-31f5-4a95-9eb1-3f61d8891ace","Type":"ContainerStarted","Data":"6a1fc546fd838cd78247e0d09aa61514319e92f549379425ab226e2584dbc32e"} Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.811901 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" podUID="659a6e11-31f5-4a95-9eb1-3f61d8891ace" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.812073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" event={"ID":"35b344ad-2ef7-48e3-8643-e93e974fb326","Type":"ContainerStarted","Data":"ed65ba3f39be2d4108252bbb293cc20965e7f65e012d5a2c8d2593127401c814"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.812100 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" event={"ID":"35b344ad-2ef7-48e3-8643-e93e974fb326","Type":"ContainerStarted","Data":"02079871637e0e454519b4a31fc5dddc6da7a263a1ca0d9ea4a1093fd4f20d4c"} Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.812608 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" podUID="fb36bed3-c7f6-43ad-b250-da42097b673e" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.814413 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" event={"ID":"45317f9d-c1aa-4f42-b52f-f910a7082d1c","Type":"ContainerStarted","Data":"d683481ee3d07d70763c6e07a701b948fa65ebb2899aa2a2f2d65cd3d60102ba"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.816082 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" event={"ID":"cefed49a-2a06-49a6-917e-ff05d6457bca","Type":"ContainerStarted","Data":"2e6442d213d4a7bf090e23a8c08c7a7fb56d092c22f2bd5375c6855b5150eb2c"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.820902 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" event={"ID":"d039dde6-16e1-42dd-a19f-b564a3939837","Type":"ContainerStarted","Data":"7cca5fe4ac75c3b09ff8eeddbd43cf0cda30d22dc84fb0c7f6b852b074b66744"} Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.821321 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" podUID="35b344ad-2ef7-48e3-8643-e93e974fb326" Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.835061 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" event={"ID":"09117b6f-d508-460f-8848-ad0ba4f7c1be","Type":"ContainerStarted","Data":"d9c79b887460e0d481c0975d10f96ad3df548046eade0336c3e7b3bc7506f013"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.837522 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" event={"ID":"2d16f135-1997-4bb6-9cb3-dae655640a01","Type":"ContainerStarted","Data":"4a79cc19fca0c64299db2dce9d24dad165dd90b8eaf9b891a06f653944ce8958"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.838920 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" event={"ID":"0d0347d8-c6bc-4bd5-b75a-28ba94906a6d","Type":"ContainerStarted","Data":"4243579fe681bf7e2da311571473d687609e181b48f9792e1b6a8ad471228f82"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.842376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" event={"ID":"11159753-a963-44c9-bb63-53340a200de7","Type":"ContainerStarted","Data":"9051ce1ea105925f9455af3a6b1b45d2a21cd000a860e9ba129faaac0a9bbd35"} Oct 07 21:10:20 crc kubenswrapper[4926]: I1007 21:10:20.842403 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" event={"ID":"11159753-a963-44c9-bb63-53340a200de7","Type":"ContainerStarted","Data":"0025e0cddabfa138d04e1ce88c0684fc2aa49d5e013f2130e6de9783d084ee5e"} Oct 07 21:10:20 crc kubenswrapper[4926]: E1007 21:10:20.843773 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" podUID="11159753-a963-44c9-bb63-53340a200de7" Oct 07 21:10:21 crc kubenswrapper[4926]: E1007 21:10:21.873582 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" podUID="f825a6cc-9414-44f5-98a2-df2c3f1e825c" Oct 07 21:10:21 crc kubenswrapper[4926]: E1007 21:10:21.873639 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" podUID="659a6e11-31f5-4a95-9eb1-3f61d8891ace" Oct 07 21:10:21 crc kubenswrapper[4926]: E1007 21:10:21.873693 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" podUID="fb36bed3-c7f6-43ad-b250-da42097b673e" Oct 07 21:10:21 crc kubenswrapper[4926]: E1007 21:10:21.873727 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" podUID="35b344ad-2ef7-48e3-8643-e93e974fb326" Oct 07 21:10:21 crc kubenswrapper[4926]: E1007 21:10:21.873759 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" podUID="11159753-a963-44c9-bb63-53340a200de7" Oct 07 21:10:32 crc kubenswrapper[4926]: E1007 21:10:32.045029 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/openstack-k8s-operators/watcher-operator:50fcb74e8b560a52907eb4f836436c06bd655091" Oct 07 21:10:32 crc kubenswrapper[4926]: E1007 21:10:32.045864 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/openstack-k8s-operators/watcher-operator:50fcb74e8b560a52907eb4f836436c06bd655091" Oct 07 21:10:32 crc kubenswrapper[4926]: E1007 21:10:32.046077 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.12:5001/openstack-k8s-operators/watcher-operator:50fcb74e8b560a52907eb4f836436c06bd655091,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cdzbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-77fdd5cc9b-q8mkp_openstack-operators(45317f9d-c1aa-4f42-b52f-f910a7082d1c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:10:32 crc kubenswrapper[4926]: E1007 21:10:32.702584 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" podUID="45317f9d-c1aa-4f42-b52f-f910a7082d1c" Oct 07 21:10:32 crc kubenswrapper[4926]: I1007 21:10:32.990417 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" event={"ID":"fc46d600-5c99-4ce4-9beb-06376dbacd8a","Type":"ContainerStarted","Data":"3606fc20307a054bf8c30b4672d603a481bc016414ca37f00882db6c7dc1bacf"} Oct 07 21:10:32 crc kubenswrapper[4926]: I1007 21:10:32.998206 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" event={"ID":"45317f9d-c1aa-4f42-b52f-f910a7082d1c","Type":"ContainerStarted","Data":"313a52f6f22fc4ba87912a5c60f0783c7499a5c3bea7ef37437ac9a178efcda6"} Oct 07 21:10:32 crc kubenswrapper[4926]: E1007 21:10:32.999376 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/openstack-k8s-operators/watcher-operator:50fcb74e8b560a52907eb4f836436c06bd655091\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" podUID="45317f9d-c1aa-4f42-b52f-f910a7082d1c" Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.000323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" event={"ID":"d1f7536d-1b34-4d39-9249-b9959aa2a00f","Type":"ContainerStarted","Data":"5dffa32ef4c7210dc16410120205b3814aa309c88dca4f90b57b811939877b0a"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.004793 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" event={"ID":"d039dde6-16e1-42dd-a19f-b564a3939837","Type":"ContainerStarted","Data":"439dd8db4d083581247264e1cbed1ab69f0ceaabf4d5d4e996e57a920622acd4"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.006387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" event={"ID":"09117b6f-d508-460f-8848-ad0ba4f7c1be","Type":"ContainerStarted","Data":"9e86e384d7107dcb2bf58159674f65db5782e1e122ca9ec0ceb16d453821f0f9"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.014992 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" event={"ID":"ce47a525-41f3-4beb-b9b3-28d85831fdda","Type":"ContainerStarted","Data":"f5c793b5c72deef1969787157ae70c67eb5fd1a37445d087bcceac9b13fa488d"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.028589 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" event={"ID":"a2ab6887-9ac0-4beb-8da0-8cb27da2ff66","Type":"ContainerStarted","Data":"7b3fc8abb78c5c362e1e4e0aef06e5fee912bc9a958cef26773b3bc536fc661b"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.033049 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" event={"ID":"2d16f135-1997-4bb6-9cb3-dae655640a01","Type":"ContainerStarted","Data":"db201cc6fc21b071d3fe138015d76e9995277a3f8ddca30519329ac09e3452d5"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.040612 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" event={"ID":"6cb125b4-a707-4506-80fc-76bc981f1560","Type":"ContainerStarted","Data":"f900dd6f08c03dc11cd5757518d061efa2708f44bcb020151693be5e91decec3"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.044452 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" event={"ID":"fea578e9-bbed-446d-beb9-5fcab1884b60","Type":"ContainerStarted","Data":"17bfbc0e604196a3cecd65fc6f0d5c18fce184f7c5a8702a68e518ae7cb0a6d4"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.051905 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" event={"ID":"bd7d6078-1c87-459e-b83e-c408537f82a1","Type":"ContainerStarted","Data":"358dbd848df222a1aa97e17830dd13b72637ebeb5e0143db404dbcd6db83be67"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.056538 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" event={"ID":"78f2ab6d-0297-49e4-91ff-ae7c8251fb0e","Type":"ContainerStarted","Data":"2f3d9e34bfc02737b25fac9848b07f8eb1d652c8fc16a6569a99938839667e5b"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.056756 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.062300 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" event={"ID":"0d0347d8-c6bc-4bd5-b75a-28ba94906a6d","Type":"ContainerStarted","Data":"984b140a2b563e94ab5ac09b20952cc2ff424320bd22dd99d2bc5880c808aa51"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.065620 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" event={"ID":"313c922a-ffe2-4d92-a221-bde117d0f196","Type":"ContainerStarted","Data":"067cf8662e98751720f323ae02be0ba7441ad5def2e60c403792001b3370e3dc"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.065902 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.068482 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" event={"ID":"bd230ff4-3566-43dc-be73-46cfe038bd76","Type":"ContainerStarted","Data":"cfb68f7a396fe92735933fd9164628d46667dc00349c2bcf26353beae336537e"} Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.209745 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.209800 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:10:33 crc kubenswrapper[4926]: I1007 21:10:33.259057 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-9554d4fbd-2cvhj" podStartSLOduration=15.259034529000001 podStartE2EDuration="15.259034529s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:10:33.252551371 +0000 UTC m=+883.291132531" watchObservedRunningTime="2025-10-07 21:10:33.259034529 +0000 UTC m=+883.297615679" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.081739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" event={"ID":"d039dde6-16e1-42dd-a19f-b564a3939837","Type":"ContainerStarted","Data":"792a1d5b2bd14c05f26e523a910643a1ce02ca1019be9f9ce96805156a0d49c4"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.084235 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" event={"ID":"f490b56a-2c4e-49ea-93fc-f2d8fb9454fd","Type":"ContainerStarted","Data":"e4a3b8281ba3ad6860a462fbc01a748eb91ac3e7f3110019a7705c13f95a0926"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.084410 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.086504 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" event={"ID":"5517a3d5-4f45-44db-9552-fcd8de7356bf","Type":"ContainerStarted","Data":"62f626b25ea0429a3a926ee6404a510d93bd25e09441fc35650c094ff4eea4a3"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.086613 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" event={"ID":"5517a3d5-4f45-44db-9552-fcd8de7356bf","Type":"ContainerStarted","Data":"c08deb0badb84eeb111ecb0859fbdbc36a3e78d45a03dbc5b0bb7d67a8360889"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.087252 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.089245 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" event={"ID":"bd230ff4-3566-43dc-be73-46cfe038bd76","Type":"ContainerStarted","Data":"aa148d29297c90170ab04c7aae2572897fa45651ae4c3ce2d61f8247dcd9cc1f"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.089662 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.091370 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" event={"ID":"bd7d6078-1c87-459e-b83e-c408537f82a1","Type":"ContainerStarted","Data":"b0294c2dcf6537d0ab868a215b7e8d5c518744cae3207c6b1d0a8b3b08998e99"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.091777 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.093355 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" event={"ID":"fea578e9-bbed-446d-beb9-5fcab1884b60","Type":"ContainerStarted","Data":"f2522c90231401021fd10ffaaa13c798904138394f0f4c23bd5e9d78924e839c"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.093631 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.095054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" event={"ID":"cefed49a-2a06-49a6-917e-ff05d6457bca","Type":"ContainerStarted","Data":"567fb59dbfa5dc7c9c66a135a944da00c0bbc394fc5204c68f9d20dd7c4c76ff"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.095384 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.096870 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" event={"ID":"d1f7536d-1b34-4d39-9249-b9959aa2a00f","Type":"ContainerStarted","Data":"552ac34ea9c8300e7990bc02ae8cf5f4c6becbf2c8886a4a34b94f894a6804a8"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.097305 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.099006 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" event={"ID":"a2ab6887-9ac0-4beb-8da0-8cb27da2ff66","Type":"ContainerStarted","Data":"8db78d861757645c654147064fde61369eb7d8645f16048c0c04fc5c4f2d3692"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.099244 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.100539 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" event={"ID":"09117b6f-d508-460f-8848-ad0ba4f7c1be","Type":"ContainerStarted","Data":"0bde5ef5b849667c0b663887cf4ddb77360badfeee917313b6343776fc1d28c5"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.100771 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.102119 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" event={"ID":"ce47a525-41f3-4beb-b9b3-28d85831fdda","Type":"ContainerStarted","Data":"39d6bbbbf25a831dbca1381b792a3539bc48eb45ac6f7dd74a1ca5bf6f0e30e1"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.102355 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.104921 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" event={"ID":"fc46d600-5c99-4ce4-9beb-06376dbacd8a","Type":"ContainerStarted","Data":"2a10c62c4d9b707004d30e9dc48d1f1177071290aefedbfdb75eee86952b9739"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.105067 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.106542 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" event={"ID":"2d16f135-1997-4bb6-9cb3-dae655640a01","Type":"ContainerStarted","Data":"8e9539bb1b941a4a725a4cf8bb40005f51e3c63e9898a5729e215deb14d6edff"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.106861 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.108063 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" podStartSLOduration=4.044019503 podStartE2EDuration="16.108053606s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:20.038345278 +0000 UTC m=+870.076926428" lastFinishedPulling="2025-10-07 21:10:32.102379381 +0000 UTC m=+882.140960531" observedRunningTime="2025-10-07 21:10:34.106013857 +0000 UTC m=+884.144595007" watchObservedRunningTime="2025-10-07 21:10:34.108053606 +0000 UTC m=+884.146634756" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.111145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" event={"ID":"6cb125b4-a707-4506-80fc-76bc981f1560","Type":"ContainerStarted","Data":"21d28689796ad2ef05aa6a567a6e42facb3f086b4bad07028aacd9bd74495626"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.111472 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.114926 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" event={"ID":"0d0347d8-c6bc-4bd5-b75a-28ba94906a6d","Type":"ContainerStarted","Data":"3b3274eaf9653f51a05c8038f2802d8fb60bd96d3fe6c97c4b1742a54cb22b29"} Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.115062 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.117653 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" event={"ID":"313c922a-ffe2-4d92-a221-bde117d0f196","Type":"ContainerStarted","Data":"69a0887a5490cab633b735c915557242bb5436aa2330210d27696d57d1a0360f"} Oct 07 21:10:34 crc kubenswrapper[4926]: E1007 21:10:34.121045 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/openstack-k8s-operators/watcher-operator:50fcb74e8b560a52907eb4f836436c06bd655091\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" podUID="45317f9d-c1aa-4f42-b52f-f910a7082d1c" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.157660 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" podStartSLOduration=4.030437591 podStartE2EDuration="16.157643331s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.959166947 +0000 UTC m=+869.997748097" lastFinishedPulling="2025-10-07 21:10:32.086372687 +0000 UTC m=+882.124953837" observedRunningTime="2025-10-07 21:10:34.138001852 +0000 UTC m=+884.176583002" watchObservedRunningTime="2025-10-07 21:10:34.157643331 +0000 UTC m=+884.196224491" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.160917 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" podStartSLOduration=5.089893241 podStartE2EDuration="17.160908605s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.962886155 +0000 UTC m=+870.001467295" lastFinishedPulling="2025-10-07 21:10:32.033901509 +0000 UTC m=+882.072482659" observedRunningTime="2025-10-07 21:10:34.157019643 +0000 UTC m=+884.195600793" watchObservedRunningTime="2025-10-07 21:10:34.160908605 +0000 UTC m=+884.199489745" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.178442 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" podStartSLOduration=4.049468095 podStartE2EDuration="17.178430222s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:18.815317689 +0000 UTC m=+868.853898839" lastFinishedPulling="2025-10-07 21:10:31.944279766 +0000 UTC m=+881.982860966" observedRunningTime="2025-10-07 21:10:34.174717755 +0000 UTC m=+884.213298905" watchObservedRunningTime="2025-10-07 21:10:34.178430222 +0000 UTC m=+884.217011372" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.192096 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" podStartSLOduration=4.669525437 podStartE2EDuration="17.192084777s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.509952539 +0000 UTC m=+869.548533689" lastFinishedPulling="2025-10-07 21:10:32.032511879 +0000 UTC m=+882.071093029" observedRunningTime="2025-10-07 21:10:34.190641976 +0000 UTC m=+884.229223126" watchObservedRunningTime="2025-10-07 21:10:34.192084777 +0000 UTC m=+884.230665927" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.207688 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" podStartSLOduration=4.640616901 podStartE2EDuration="17.207679179s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.377238208 +0000 UTC m=+869.415819358" lastFinishedPulling="2025-10-07 21:10:31.944300446 +0000 UTC m=+881.982881636" observedRunningTime="2025-10-07 21:10:34.206298779 +0000 UTC m=+884.244879929" watchObservedRunningTime="2025-10-07 21:10:34.207679179 +0000 UTC m=+884.246260329" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.234323 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" podStartSLOduration=4.325572444 podStartE2EDuration="17.234307029s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.035548951 +0000 UTC m=+869.074130101" lastFinishedPulling="2025-10-07 21:10:31.944283536 +0000 UTC m=+881.982864686" observedRunningTime="2025-10-07 21:10:34.228330476 +0000 UTC m=+884.266911626" watchObservedRunningTime="2025-10-07 21:10:34.234307029 +0000 UTC m=+884.272888179" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.249236 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" podStartSLOduration=4.657392256 podStartE2EDuration="17.24918331s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.476343366 +0000 UTC m=+869.514924516" lastFinishedPulling="2025-10-07 21:10:32.06813442 +0000 UTC m=+882.106715570" observedRunningTime="2025-10-07 21:10:34.245029399 +0000 UTC m=+884.283610549" watchObservedRunningTime="2025-10-07 21:10:34.24918331 +0000 UTC m=+884.287764460" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.272480 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" podStartSLOduration=5.13858824 podStartE2EDuration="17.272467393s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.965249073 +0000 UTC m=+870.003830223" lastFinishedPulling="2025-10-07 21:10:32.099128226 +0000 UTC m=+882.137709376" observedRunningTime="2025-10-07 21:10:34.269216179 +0000 UTC m=+884.307797329" watchObservedRunningTime="2025-10-07 21:10:34.272467393 +0000 UTC m=+884.311048543" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.291884 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" podStartSLOduration=4.664695777 podStartE2EDuration="17.291873315s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.405537477 +0000 UTC m=+869.444118627" lastFinishedPulling="2025-10-07 21:10:32.032715015 +0000 UTC m=+882.071296165" observedRunningTime="2025-10-07 21:10:34.289069054 +0000 UTC m=+884.327650204" watchObservedRunningTime="2025-10-07 21:10:34.291873315 +0000 UTC m=+884.330454465" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.343232 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" podStartSLOduration=4.688397483 podStartE2EDuration="17.343216321s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.377690751 +0000 UTC m=+869.416271901" lastFinishedPulling="2025-10-07 21:10:32.032509559 +0000 UTC m=+882.071090739" observedRunningTime="2025-10-07 21:10:34.31485125 +0000 UTC m=+884.353432400" watchObservedRunningTime="2025-10-07 21:10:34.343216321 +0000 UTC m=+884.381797471" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.367013 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" podStartSLOduration=4.325322717 podStartE2EDuration="17.366996559s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:18.991654521 +0000 UTC m=+869.030235671" lastFinishedPulling="2025-10-07 21:10:32.033328363 +0000 UTC m=+882.071909513" observedRunningTime="2025-10-07 21:10:34.365836035 +0000 UTC m=+884.404417185" watchObservedRunningTime="2025-10-07 21:10:34.366996559 +0000 UTC m=+884.405577709" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.415504 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" podStartSLOduration=4.330842623 podStartE2EDuration="16.415489162s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.950522887 +0000 UTC m=+869.989104037" lastFinishedPulling="2025-10-07 21:10:32.035169426 +0000 UTC m=+882.073750576" observedRunningTime="2025-10-07 21:10:34.391161548 +0000 UTC m=+884.429742688" watchObservedRunningTime="2025-10-07 21:10:34.415489162 +0000 UTC m=+884.454070302" Oct 07 21:10:34 crc kubenswrapper[4926]: I1007 21:10:34.416880 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" podStartSLOduration=4.320298648 podStartE2EDuration="16.416873072s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.950876777 +0000 UTC m=+869.989457927" lastFinishedPulling="2025-10-07 21:10:32.047451201 +0000 UTC m=+882.086032351" observedRunningTime="2025-10-07 21:10:34.413524565 +0000 UTC m=+884.452105715" watchObservedRunningTime="2025-10-07 21:10:34.416873072 +0000 UTC m=+884.455454222" Oct 07 21:10:35 crc kubenswrapper[4926]: I1007 21:10:35.133417 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" event={"ID":"f490b56a-2c4e-49ea-93fc-f2d8fb9454fd","Type":"ContainerStarted","Data":"63908ccfb290c2b7363860d4437722c686018f6bb276918f23f2d7f4ad504df0"} Oct 07 21:10:35 crc kubenswrapper[4926]: I1007 21:10:35.137978 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" event={"ID":"cefed49a-2a06-49a6-917e-ff05d6457bca","Type":"ContainerStarted","Data":"1541733716c74a1adc4a64bd8d9871ae0d6c9d21f96742d7fc5e058bf1c12a79"} Oct 07 21:10:35 crc kubenswrapper[4926]: I1007 21:10:35.140044 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:35 crc kubenswrapper[4926]: I1007 21:10:35.157468 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" podStartSLOduration=4.982267482 podStartE2EDuration="17.157452381s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.923716781 +0000 UTC m=+869.962297931" lastFinishedPulling="2025-10-07 21:10:32.09890168 +0000 UTC m=+882.137482830" observedRunningTime="2025-10-07 21:10:35.156650278 +0000 UTC m=+885.195231458" watchObservedRunningTime="2025-10-07 21:10:35.157452381 +0000 UTC m=+885.196033531" Oct 07 21:10:35 crc kubenswrapper[4926]: I1007 21:10:35.160883 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" podStartSLOduration=5.547392279 podStartE2EDuration="18.16085746s" podCreationTimestamp="2025-10-07 21:10:17 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.449488699 +0000 UTC m=+869.488069849" lastFinishedPulling="2025-10-07 21:10:32.06295388 +0000 UTC m=+882.101535030" observedRunningTime="2025-10-07 21:10:34.440438224 +0000 UTC m=+884.479019384" watchObservedRunningTime="2025-10-07 21:10:35.16085746 +0000 UTC m=+885.199438610" Oct 07 21:10:36 crc kubenswrapper[4926]: I1007 21:10:36.147154 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.155933 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" event={"ID":"35b344ad-2ef7-48e3-8643-e93e974fb326","Type":"ContainerStarted","Data":"92953e590e9badb46b3a2672a6e139e01bbc6c0bc60d0cbb68d0928359944e37"} Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.156483 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.158482 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" event={"ID":"11159753-a963-44c9-bb63-53340a200de7","Type":"ContainerStarted","Data":"02147716e071402d938ed595890c3242547ff6d5c548d737472e40f7f693750b"} Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.158698 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.161150 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" event={"ID":"f825a6cc-9414-44f5-98a2-df2c3f1e825c","Type":"ContainerStarted","Data":"1ac2bb1abcd645922d0cc0f39d3b43fbb80ea9d972b87608af7c2dcd0eab44fb"} Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.167858 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-47m5t" Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.175534 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" podStartSLOduration=2.286056926 podStartE2EDuration="19.175514555s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:19.975409677 +0000 UTC m=+870.013990827" lastFinishedPulling="2025-10-07 21:10:36.864867306 +0000 UTC m=+886.903448456" observedRunningTime="2025-10-07 21:10:37.17085801 +0000 UTC m=+887.209439160" watchObservedRunningTime="2025-10-07 21:10:37.175514555 +0000 UTC m=+887.214095705" Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.194335 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" podStartSLOduration=2.381168188 podStartE2EDuration="19.194312959s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:20.046732021 +0000 UTC m=+870.085313171" lastFinishedPulling="2025-10-07 21:10:36.859876802 +0000 UTC m=+886.898457942" observedRunningTime="2025-10-07 21:10:37.186461402 +0000 UTC m=+887.225042562" watchObservedRunningTime="2025-10-07 21:10:37.194312959 +0000 UTC m=+887.232894119" Oct 07 21:10:37 crc kubenswrapper[4926]: I1007 21:10:37.244278 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" podStartSLOduration=2.671051166 podStartE2EDuration="19.244257274s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:20.312187642 +0000 UTC m=+870.350768792" lastFinishedPulling="2025-10-07 21:10:36.88539375 +0000 UTC m=+886.923974900" observedRunningTime="2025-10-07 21:10:37.239345732 +0000 UTC m=+887.277926892" watchObservedRunningTime="2025-10-07 21:10:37.244257274 +0000 UTC m=+887.282838424" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.154731 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-7ltns" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.171795 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-bljj9" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.200885 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-pwsh8" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.226801 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-jnbmb" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.231578 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-mwlfv" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.256339 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-nm2mk" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.282841 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cgkzf" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.336461 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-649675d675-zz2jk" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.345681 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-tmh94" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.446670 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-wxrzg" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.556111 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-7spjf" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.587178 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-tjfcp" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.621980 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4xz96" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.652414 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-rxdc8" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.703837 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-tsklh" Oct 07 21:10:38 crc kubenswrapper[4926]: I1007 21:10:38.830260 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:39 crc kubenswrapper[4926]: I1007 21:10:39.185575 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" event={"ID":"659a6e11-31f5-4a95-9eb1-3f61d8891ace","Type":"ContainerStarted","Data":"05dd41bdfa9733f0f1332031780341a8b1d966100c5664dcfeb97220ada91ba3"} Oct 07 21:10:39 crc kubenswrapper[4926]: I1007 21:10:39.185773 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:39 crc kubenswrapper[4926]: I1007 21:10:39.204176 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" podStartSLOduration=3.017548993 podStartE2EDuration="21.204156276s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:20.050802069 +0000 UTC m=+870.089383229" lastFinishedPulling="2025-10-07 21:10:38.237409362 +0000 UTC m=+888.275990512" observedRunningTime="2025-10-07 21:10:39.199746718 +0000 UTC m=+889.238327878" watchObservedRunningTime="2025-10-07 21:10:39.204156276 +0000 UTC m=+889.242737426" Oct 07 21:10:40 crc kubenswrapper[4926]: I1007 21:10:40.197571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" event={"ID":"fb36bed3-c7f6-43ad-b250-da42097b673e","Type":"ContainerStarted","Data":"a57f76deae5c6b14f5b0071638e7b65495e8663ad5e3f5a5664c2325c7a1e31b"} Oct 07 21:10:40 crc kubenswrapper[4926]: I1007 21:10:40.222244 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5" podStartSLOduration=2.955144105 podStartE2EDuration="22.222228574s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:20.313546651 +0000 UTC m=+870.352127801" lastFinishedPulling="2025-10-07 21:10:39.58063112 +0000 UTC m=+889.619212270" observedRunningTime="2025-10-07 21:10:40.218805425 +0000 UTC m=+890.257386595" watchObservedRunningTime="2025-10-07 21:10:40.222228574 +0000 UTC m=+890.260809724" Oct 07 21:10:47 crc kubenswrapper[4926]: I1007 21:10:47.261829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" event={"ID":"45317f9d-c1aa-4f42-b52f-f910a7082d1c","Type":"ContainerStarted","Data":"e125c3dc08f0dcf3522ce38d1b80c5159ef118558feb71f84071a68a53013520"} Oct 07 21:10:47 crc kubenswrapper[4926]: I1007 21:10:47.262805 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:10:47 crc kubenswrapper[4926]: I1007 21:10:47.283463 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" podStartSLOduration=2.4916639050000002 podStartE2EDuration="29.283444494s" podCreationTimestamp="2025-10-07 21:10:18 +0000 UTC" firstStartedPulling="2025-10-07 21:10:20.294408598 +0000 UTC m=+870.332989748" lastFinishedPulling="2025-10-07 21:10:47.086189177 +0000 UTC m=+897.124770337" observedRunningTime="2025-10-07 21:10:47.278935073 +0000 UTC m=+897.317516243" watchObservedRunningTime="2025-10-07 21:10:47.283444494 +0000 UTC m=+897.322025654" Oct 07 21:10:48 crc kubenswrapper[4926]: I1007 21:10:48.720007 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-59xh2" Oct 07 21:10:48 crc kubenswrapper[4926]: I1007 21:10:48.801371 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-w2bgf" Oct 07 21:10:48 crc kubenswrapper[4926]: I1007 21:10:48.834300 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-rtwpd" Oct 07 21:10:49 crc kubenswrapper[4926]: I1007 21:10:49.282692 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc" Oct 07 21:10:58 crc kubenswrapper[4926]: I1007 21:10:58.990085 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-77fdd5cc9b-q8mkp" Oct 07 21:11:03 crc kubenswrapper[4926]: I1007 21:11:03.209947 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:11:03 crc kubenswrapper[4926]: I1007 21:11:03.210382 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:11:17 crc kubenswrapper[4926]: I1007 21:11:17.919435 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d9454c9c9-z28fx"] Oct 07 21:11:17 crc kubenswrapper[4926]: I1007 21:11:17.920985 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:17 crc kubenswrapper[4926]: W1007 21:11:17.925491 4926 reflector.go:561] object-"openstack"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 07 21:11:17 crc kubenswrapper[4926]: E1007 21:11:17.925537 4926 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 21:11:17 crc kubenswrapper[4926]: W1007 21:11:17.928364 4926 reflector.go:561] object-"openstack"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 07 21:11:17 crc kubenswrapper[4926]: W1007 21:11:17.928379 4926 reflector.go:561] object-"openstack"/"dnsmasq-dns-dockercfg-gk595": failed to list *v1.Secret: secrets "dnsmasq-dns-dockercfg-gk595" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 07 21:11:17 crc kubenswrapper[4926]: E1007 21:11:17.928413 4926 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 21:11:17 crc kubenswrapper[4926]: W1007 21:11:17.928427 4926 reflector.go:561] object-"openstack"/"dns": failed to list *v1.ConfigMap: configmaps "dns" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 07 21:11:17 crc kubenswrapper[4926]: E1007 21:11:17.928435 4926 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"dnsmasq-dns-dockercfg-gk595\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"dnsmasq-dns-dockercfg-gk595\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 21:11:17 crc kubenswrapper[4926]: E1007 21:11:17.928480 4926 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"dns\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"dns\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 21:11:17 crc kubenswrapper[4926]: I1007 21:11:17.933321 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9454c9c9-z28fx"] Oct 07 21:11:17 crc kubenswrapper[4926]: I1007 21:11:17.972912 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-config\") pod \"dnsmasq-dns-6d9454c9c9-z28fx\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:17 crc kubenswrapper[4926]: I1007 21:11:17.973022 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hr4v\" (UniqueName: \"kubernetes.io/projected/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-kube-api-access-4hr4v\") pod \"dnsmasq-dns-6d9454c9c9-z28fx\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.004031 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567d68c74f-df6kp"] Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.005227 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.007236 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.012098 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567d68c74f-df6kp"] Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.074738 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-config\") pod \"dnsmasq-dns-6d9454c9c9-z28fx\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.075038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-config\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.075126 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hr4v\" (UniqueName: \"kubernetes.io/projected/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-kube-api-access-4hr4v\") pod \"dnsmasq-dns-6d9454c9c9-z28fx\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.075153 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-dns-svc\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.075177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptbrp\" (UniqueName: \"kubernetes.io/projected/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-kube-api-access-ptbrp\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.176527 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-config\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.176624 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-dns-svc\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.176656 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptbrp\" (UniqueName: \"kubernetes.io/projected/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-kube-api-access-ptbrp\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.177540 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-dns-svc\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.920107 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.925992 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.927786 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-config\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:18 crc kubenswrapper[4926]: I1007 21:11:18.936653 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-config\") pod \"dnsmasq-dns-6d9454c9c9-z28fx\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.108376 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gk595" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.308607 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.319271 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptbrp\" (UniqueName: \"kubernetes.io/projected/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-kube-api-access-ptbrp\") pod \"dnsmasq-dns-567d68c74f-df6kp\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.320465 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hr4v\" (UniqueName: \"kubernetes.io/projected/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-kube-api-access-4hr4v\") pod \"dnsmasq-dns-6d9454c9c9-z28fx\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.443320 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.533788 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.745633 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9454c9c9-z28fx"] Oct 07 21:11:19 crc kubenswrapper[4926]: I1007 21:11:19.875110 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567d68c74f-df6kp"] Oct 07 21:11:19 crc kubenswrapper[4926]: W1007 21:11:19.884080 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb1ff3fe_9ce5_455c_a414_a3c3a042bd39.slice/crio-1c4893cacf23d4c0b89c6dde170d81c477c025dc66f0854420568b399a56da7a WatchSource:0}: Error finding container 1c4893cacf23d4c0b89c6dde170d81c477c025dc66f0854420568b399a56da7a: Status 404 returned error can't find the container with id 1c4893cacf23d4c0b89c6dde170d81c477c025dc66f0854420568b399a56da7a Oct 07 21:11:20 crc kubenswrapper[4926]: I1007 21:11:20.572124 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" event={"ID":"2c31b57b-e4b4-4a8f-945a-ef221de9dec4","Type":"ContainerStarted","Data":"11083f4b35787ab3e43d1fc97df3597515cf00327f5ec9b92c522b32c17dc4c1"} Oct 07 21:11:20 crc kubenswrapper[4926]: I1007 21:11:20.574322 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" event={"ID":"db1ff3fe-9ce5-455c-a414-a3c3a042bd39","Type":"ContainerStarted","Data":"1c4893cacf23d4c0b89c6dde170d81c477c025dc66f0854420568b399a56da7a"} Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.046328 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567d68c74f-df6kp"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.074618 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c64b7c769-75lt2"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.079188 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.102315 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c64b7c769-75lt2"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.251465 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcv4f\" (UniqueName: \"kubernetes.io/projected/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-kube-api-access-zcv4f\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.251593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-config\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.251614 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-dns-svc\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.336850 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9454c9c9-z28fx"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.353304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcv4f\" (UniqueName: \"kubernetes.io/projected/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-kube-api-access-zcv4f\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.353436 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-config\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.353465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-dns-svc\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.354378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-dns-svc\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.355713 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-config\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.385216 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5544966479-schqv"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.386795 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.397674 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcv4f\" (UniqueName: \"kubernetes.io/projected/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-kube-api-access-zcv4f\") pod \"dnsmasq-dns-5c64b7c769-75lt2\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.406627 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.428284 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5544966479-schqv"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.557982 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-config\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.558038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mll45\" (UniqueName: \"kubernetes.io/projected/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-kube-api-access-mll45\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.558088 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-dns-svc\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.666227 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mll45\" (UniqueName: \"kubernetes.io/projected/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-kube-api-access-mll45\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.666324 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-dns-svc\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.666468 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-config\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.667332 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-config\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.667829 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-dns-svc\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.693215 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mll45\" (UniqueName: \"kubernetes.io/projected/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-kube-api-access-mll45\") pod \"dnsmasq-dns-5544966479-schqv\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.749965 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.777071 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5544966479-schqv"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.817767 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54b6445897-mtczn"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.819381 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.825307 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b6445897-mtczn"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.892882 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c64b7c769-75lt2"] Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.975831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-dns-svc\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.975873 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76lls\" (UniqueName: \"kubernetes.io/projected/76602401-809f-4d0f-8683-171a94389640-kube-api-access-76lls\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:22 crc kubenswrapper[4926]: I1007 21:11:22.975944 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-config\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.078406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-config\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.078563 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-dns-svc\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.078596 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76lls\" (UniqueName: \"kubernetes.io/projected/76602401-809f-4d0f-8683-171a94389640-kube-api-access-76lls\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.079532 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-dns-svc\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.079723 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-config\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.098907 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76lls\" (UniqueName: \"kubernetes.io/projected/76602401-809f-4d0f-8683-171a94389640-kube-api-access-76lls\") pod \"dnsmasq-dns-54b6445897-mtczn\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.142051 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.249120 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.250464 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.254141 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.254274 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.254279 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.257649 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lmqmr" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.257855 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.257871 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.258013 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.263239 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.319977 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5544966479-schqv"] Oct 07 21:11:23 crc kubenswrapper[4926]: W1007 21:11:23.370925 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd12e9cf1_65dc_4c7c_a563_35a76b6fd1bc.slice/crio-608448ab46333326d99c24420340b161d0fdea32f984711ed5b4fc885e71da5c WatchSource:0}: Error finding container 608448ab46333326d99c24420340b161d0fdea32f984711ed5b4fc885e71da5c: Status 404 returned error can't find the container with id 608448ab46333326d99c24420340b161d0fdea32f984711ed5b4fc885e71da5c Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385767 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385836 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385899 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385921 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385950 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385980 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.385996 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.386014 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.386034 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5lls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-kube-api-access-z5lls\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.386058 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-config-data\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487302 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487346 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5lls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-kube-api-access-z5lls\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487381 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-config-data\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487440 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487478 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487536 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487555 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487574 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487603 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487624 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.487995 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.488095 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.488218 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.488224 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.490395 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-config-data\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.491925 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.492157 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.493510 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.497472 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.502526 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5lls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-kube-api-access-z5lls\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.504979 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.516864 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.527701 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.528941 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.532330 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.532517 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.532632 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.533042 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5k582" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.533251 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.533574 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.533812 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.553029 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.583748 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.623608 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5544966479-schqv" event={"ID":"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc","Type":"ContainerStarted","Data":"608448ab46333326d99c24420340b161d0fdea32f984711ed5b4fc885e71da5c"} Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.625972 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" event={"ID":"4d47eec2-94cc-47d5-90fe-a3d17b72ec99","Type":"ContainerStarted","Data":"4aa4674218b8e29e9d4dd2eead33f7b29f9ff6f2b0bd67590cb8466a49de642c"} Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.650969 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b6445897-mtczn"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690098 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6hvx\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-kube-api-access-f6hvx\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690161 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690201 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690229 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690286 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690331 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690422 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690457 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690483 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690508 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.690532 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.793899 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795087 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795114 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795137 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795179 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6hvx\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-kube-api-access-f6hvx\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795274 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795294 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795317 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795369 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795408 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795593 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.795894 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.796269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.796444 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.796907 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.805836 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.806431 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.806707 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.810844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.811295 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.819481 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.822032 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6hvx\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-kube-api-access-f6hvx\") pod \"rabbitmq-cell1-server-0\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.887271 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.938249 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.945744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.950721 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.950744 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-cbpvs" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.950993 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.951137 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.951409 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.951456 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.951503 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Oct 07 21:11:23 crc kubenswrapper[4926]: I1007 21:11:23.951517 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099400 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099426 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099451 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099472 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5b0d7388-05f4-4df6-b093-c9157a4658b0-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099490 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5b0d7388-05f4-4df6-b093-c9157a4658b0-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099510 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099528 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099556 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq55l\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-kube-api-access-fq55l\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.099606 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.135036 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.200794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.200850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.200882 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.200912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5b0d7388-05f4-4df6-b093-c9157a4658b0-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.200941 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5b0d7388-05f4-4df6-b093-c9157a4658b0-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201529 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201565 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201601 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq55l\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-kube-api-access-fq55l\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201633 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201661 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201725 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.201822 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.202137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.202460 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.202662 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.202772 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5b0d7388-05f4-4df6-b093-c9157a4658b0-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.203750 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.207298 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5b0d7388-05f4-4df6-b093-c9157a4658b0-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.208102 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5b0d7388-05f4-4df6-b093-c9157a4658b0-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.210363 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.216583 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.219075 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq55l\" (UniqueName: \"kubernetes.io/projected/5b0d7388-05f4-4df6-b093-c9157a4658b0-kube-api-access-fq55l\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.232610 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"5b0d7388-05f4-4df6-b093-c9157a4658b0\") " pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.266423 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.356463 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:11:24 crc kubenswrapper[4926]: W1007 21:11:24.368588 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59e27b2d_9c0d_46a6_8be6_eee9cdf04c57.slice/crio-92ca6955f5c8830dfdc19799eb080626c55ec8e7d8f9ffceba836b2003e7b1ad WatchSource:0}: Error finding container 92ca6955f5c8830dfdc19799eb080626c55ec8e7d8f9ffceba836b2003e7b1ad: Status 404 returned error can't find the container with id 92ca6955f5c8830dfdc19799eb080626c55ec8e7d8f9ffceba836b2003e7b1ad Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.640029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0","Type":"ContainerStarted","Data":"c4f75ba6a03bce10213bd90cfd506bb75e16eae00d638627ad4eb4e2882a3eb1"} Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.647829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57","Type":"ContainerStarted","Data":"92ca6955f5c8830dfdc19799eb080626c55ec8e7d8f9ffceba836b2003e7b1ad"} Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.653449 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b6445897-mtczn" event={"ID":"76602401-809f-4d0f-8683-171a94389640","Type":"ContainerStarted","Data":"472a4b10f82dcf3e3b9b078d94ff8b6e5134a2889eb9f690ad0aa921010e4aa1"} Oct 07 21:11:24 crc kubenswrapper[4926]: I1007 21:11:24.764125 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 07 21:11:25 crc kubenswrapper[4926]: I1007 21:11:25.681732 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"5b0d7388-05f4-4df6-b093-c9157a4658b0","Type":"ContainerStarted","Data":"2ce31e69e531c030703b4ef4ddccab72a6f35e546a73e6f7694e4caff5377f4c"} Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.836786 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.839026 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.841914 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.841919 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.842013 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.842475 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.843606 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8x257" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.846873 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.852239 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.864576 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.873239 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.878738 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.879579 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.879811 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-n9lbd" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.880419 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 07 21:11:26 crc kubenswrapper[4926]: I1007 21:11:26.908080 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005013 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005295 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005319 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005388 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005413 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d169a79-a57d-458f-a790-3623f6dd0885-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spdwp\" (UniqueName: \"kubernetes.io/projected/2d169a79-a57d-458f-a790-3623f6dd0885-kube-api-access-spdwp\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005502 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005524 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005541 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005572 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005588 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-kolla-config\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005608 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005632 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-config-data-default\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r9qj\" (UniqueName: \"kubernetes.io/projected/7bf0bb92-3a4f-44ae-be03-e260cb246adb-kube-api-access-5r9qj\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005683 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005701 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-secrets\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.005721 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bf0bb92-3a4f-44ae-be03-e260cb246adb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124452 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124488 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d169a79-a57d-458f-a790-3623f6dd0885-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124513 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spdwp\" (UniqueName: \"kubernetes.io/projected/2d169a79-a57d-458f-a790-3623f6dd0885-kube-api-access-spdwp\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124538 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124564 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124582 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124642 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-kolla-config\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124667 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124695 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-config-data-default\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124744 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r9qj\" (UniqueName: \"kubernetes.io/projected/7bf0bb92-3a4f-44ae-be03-e260cb246adb-kube-api-access-5r9qj\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124770 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124786 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-secrets\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bf0bb92-3a4f-44ae-be03-e260cb246adb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124832 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.124868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.125517 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-kolla-config\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.126112 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.126639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d169a79-a57d-458f-a790-3623f6dd0885-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.126928 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bf0bb92-3a4f-44ae-be03-e260cb246adb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.127372 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-config-data-default\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.128326 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf0bb92-3a4f-44ae-be03-e260cb246adb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.130702 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.130745 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.131244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.131948 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.134243 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d169a79-a57d-458f-a790-3623f6dd0885-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.136732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-secrets\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.148675 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.149400 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf0bb92-3a4f-44ae-be03-e260cb246adb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.154258 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.171636 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spdwp\" (UniqueName: \"kubernetes.io/projected/2d169a79-a57d-458f-a790-3623f6dd0885-kube-api-access-spdwp\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.181311 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d169a79-a57d-458f-a790-3623f6dd0885-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.182441 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.192560 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r9qj\" (UniqueName: \"kubernetes.io/projected/7bf0bb92-3a4f-44ae-be03-e260cb246adb-kube-api-access-5r9qj\") pod \"openstack-galera-0\" (UID: \"7bf0bb92-3a4f-44ae-be03-e260cb246adb\") " pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.207162 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2d169a79-a57d-458f-a790-3623f6dd0885\") " pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.244380 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.250020 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.253648 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.253957 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.254211 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-slg5l" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.261474 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.330907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/038e2e94-c12a-4d54-8b48-96d08012ab97-memcached-tls-certs\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.331233 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/038e2e94-c12a-4d54-8b48-96d08012ab97-config-data\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.331307 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/038e2e94-c12a-4d54-8b48-96d08012ab97-kolla-config\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.331774 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tggsz\" (UniqueName: \"kubernetes.io/projected/038e2e94-c12a-4d54-8b48-96d08012ab97-kube-api-access-tggsz\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.331879 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/038e2e94-c12a-4d54-8b48-96d08012ab97-combined-ca-bundle\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.435167 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/038e2e94-c12a-4d54-8b48-96d08012ab97-kolla-config\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.435317 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tggsz\" (UniqueName: \"kubernetes.io/projected/038e2e94-c12a-4d54-8b48-96d08012ab97-kube-api-access-tggsz\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.435352 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/038e2e94-c12a-4d54-8b48-96d08012ab97-combined-ca-bundle\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.435383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/038e2e94-c12a-4d54-8b48-96d08012ab97-memcached-tls-certs\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.435415 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/038e2e94-c12a-4d54-8b48-96d08012ab97-config-data\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.436158 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/038e2e94-c12a-4d54-8b48-96d08012ab97-config-data\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.438081 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/038e2e94-c12a-4d54-8b48-96d08012ab97-kolla-config\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.438667 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/038e2e94-c12a-4d54-8b48-96d08012ab97-combined-ca-bundle\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.453005 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/038e2e94-c12a-4d54-8b48-96d08012ab97-memcached-tls-certs\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.456465 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tggsz\" (UniqueName: \"kubernetes.io/projected/038e2e94-c12a-4d54-8b48-96d08012ab97-kube-api-access-tggsz\") pod \"memcached-0\" (UID: \"038e2e94-c12a-4d54-8b48-96d08012ab97\") " pod="openstack/memcached-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.458597 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.509798 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 21:11:27 crc kubenswrapper[4926]: I1007 21:11:27.593954 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.127147 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.129298 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.132468 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8hllq" Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.132738 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.174695 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtk6r\" (UniqueName: \"kubernetes.io/projected/f838140a-347d-4bb3-8cc0-b4100bfe4b35-kube-api-access-wtk6r\") pod \"kube-state-metrics-0\" (UID: \"f838140a-347d-4bb3-8cc0-b4100bfe4b35\") " pod="openstack/kube-state-metrics-0" Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.276604 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtk6r\" (UniqueName: \"kubernetes.io/projected/f838140a-347d-4bb3-8cc0-b4100bfe4b35-kube-api-access-wtk6r\") pod \"kube-state-metrics-0\" (UID: \"f838140a-347d-4bb3-8cc0-b4100bfe4b35\") " pod="openstack/kube-state-metrics-0" Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.297964 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtk6r\" (UniqueName: \"kubernetes.io/projected/f838140a-347d-4bb3-8cc0-b4100bfe4b35-kube-api-access-wtk6r\") pod \"kube-state-metrics-0\" (UID: \"f838140a-347d-4bb3-8cc0-b4100bfe4b35\") " pod="openstack/kube-state-metrics-0" Oct 07 21:11:29 crc kubenswrapper[4926]: I1007 21:11:29.451284 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.434555 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.436727 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.438802 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.438966 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.439510 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hckhp" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.439804 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.447143 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.448993 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.456090 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599351 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599398 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599483 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v28n\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-kube-api-access-9v28n\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599734 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599832 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ad789bb-88f9-4bec-ba17-46ac114893aa-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599917 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.599978 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ad789bb-88f9-4bec-ba17-46ac114893aa-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.700797 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.700852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ad789bb-88f9-4bec-ba17-46ac114893aa-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.700886 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.700920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ad789bb-88f9-4bec-ba17-46ac114893aa-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.700953 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.700978 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.701016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.701075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v28n\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-kube-api-access-9v28n\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.715607 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.716072 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.716276 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ad789bb-88f9-4bec-ba17-46ac114893aa-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.720525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ad789bb-88f9-4bec-ba17-46ac114893aa-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.722170 4926 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.722252 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/981924a86d194783a644e1657994993df571041fb46e6d3f243affb52a3a59ba/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.722471 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.728008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.739623 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v28n\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-kube-api-access-9v28n\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:30 crc kubenswrapper[4926]: I1007 21:11:30.793301 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:31 crc kubenswrapper[4926]: I1007 21:11:31.060379 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.210797 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.211082 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.211131 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.211915 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a7195a5bc10effcad121b73904803864abd9373d3dda56955bffbf5b0e3d1427"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.211980 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://a7195a5bc10effcad121b73904803864abd9373d3dda56955bffbf5b0e3d1427" gracePeriod=600 Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.289139 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.290722 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.292841 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.292872 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.293843 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.294488 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.294694 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-g7fv2" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.310047 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449143 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449320 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwjfr\" (UniqueName: \"kubernetes.io/projected/b2704dcf-cd03-4f91-af4e-ea708f3c0534-kube-api-access-dwjfr\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449399 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2704dcf-cd03-4f91-af4e-ea708f3c0534-config\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449453 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2704dcf-cd03-4f91-af4e-ea708f3c0534-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449541 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b2704dcf-cd03-4f91-af4e-ea708f3c0534-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449576 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449643 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.449706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.453021 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6qkqt"] Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.454095 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.458561 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.458953 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-gztbv" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.459683 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.469180 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6qkqt"] Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.483668 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-zg8nq"] Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.486233 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.505549 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zg8nq"] Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551472 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551561 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-log-ovn\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551641 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwjfr\" (UniqueName: \"kubernetes.io/projected/b2704dcf-cd03-4f91-af4e-ea708f3c0534-kube-api-access-dwjfr\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-run\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551698 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-ovn-controller-tls-certs\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551719 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2704dcf-cd03-4f91-af4e-ea708f3c0534-config\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2704dcf-cd03-4f91-af4e-ea708f3c0534-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551773 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b2704dcf-cd03-4f91-af4e-ea708f3c0534-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551800 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551817 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-scripts\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551846 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-run-ovn\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551864 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpmvz\" (UniqueName: \"kubernetes.io/projected/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-kube-api-access-hpmvz\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551913 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-combined-ca-bundle\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.551928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.553918 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.554020 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2704dcf-cd03-4f91-af4e-ea708f3c0534-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.554287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b2704dcf-cd03-4f91-af4e-ea708f3c0534-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.555973 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2704dcf-cd03-4f91-af4e-ea708f3c0534-config\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.558091 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.567275 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.568956 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2704dcf-cd03-4f91-af4e-ea708f3c0534-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.571042 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwjfr\" (UniqueName: \"kubernetes.io/projected/b2704dcf-cd03-4f91-af4e-ea708f3c0534-kube-api-access-dwjfr\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.596381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b2704dcf-cd03-4f91-af4e-ea708f3c0534\") " pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.606595 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653719 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpmvz\" (UniqueName: \"kubernetes.io/projected/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-kube-api-access-hpmvz\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-combined-ca-bundle\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653806 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-lib\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653824 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-log-ovn\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653860 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v8gf\" (UniqueName: \"kubernetes.io/projected/188a9695-679b-46c5-ac4b-26f71c9b4110-kube-api-access-2v8gf\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653911 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-log\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653926 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-etc-ovs\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653947 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-run\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653971 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-ovn-controller-tls-certs\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.653994 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/188a9695-679b-46c5-ac4b-26f71c9b4110-scripts\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.654015 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-run\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.654044 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-scripts\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.654071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-run-ovn\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.654520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-run-ovn\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.655933 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-log-ovn\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.656071 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-var-run\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.658244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-combined-ca-bundle\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.668832 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-scripts\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.671432 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpmvz\" (UniqueName: \"kubernetes.io/projected/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-kube-api-access-hpmvz\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.682734 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2-ovn-controller-tls-certs\") pod \"ovn-controller-6qkqt\" (UID: \"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2\") " pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/188a9695-679b-46c5-ac4b-26f71c9b4110-scripts\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755427 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-run\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755548 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-lib\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755604 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v8gf\" (UniqueName: \"kubernetes.io/projected/188a9695-679b-46c5-ac4b-26f71c9b4110-kube-api-access-2v8gf\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755644 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-log\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755659 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-etc-ovs\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.755912 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-etc-ovs\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.758258 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/188a9695-679b-46c5-ac4b-26f71c9b4110-scripts\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.758321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-run\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.758484 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-lib\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.759000 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/188a9695-679b-46c5-ac4b-26f71c9b4110-var-log\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.777347 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v8gf\" (UniqueName: \"kubernetes.io/projected/188a9695-679b-46c5-ac4b-26f71c9b4110-kube-api-access-2v8gf\") pod \"ovn-controller-ovs-zg8nq\" (UID: \"188a9695-679b-46c5-ac4b-26f71c9b4110\") " pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.798830 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:33 crc kubenswrapper[4926]: I1007 21:11:33.840410 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:36 crc kubenswrapper[4926]: I1007 21:11:36.837917 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="a7195a5bc10effcad121b73904803864abd9373d3dda56955bffbf5b0e3d1427" exitCode=0 Oct 07 21:11:36 crc kubenswrapper[4926]: I1007 21:11:36.838970 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"a7195a5bc10effcad121b73904803864abd9373d3dda56955bffbf5b0e3d1427"} Oct 07 21:11:36 crc kubenswrapper[4926]: I1007 21:11:36.839081 4926 scope.go:117] "RemoveContainer" containerID="b152fe038f21276c6a7128f942977a43dfd86caa4be6d3a00a66f622f88a2161" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.023447 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.024890 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.027261 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.027282 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.027478 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7vl67" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.027657 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.040735 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110530 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110651 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110716 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdc18634-337d-4562-a6e5-a20df1f53028-config\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110777 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110811 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110846 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnm7r\" (UniqueName: \"kubernetes.io/projected/fdc18634-337d-4562-a6e5-a20df1f53028-kube-api-access-qnm7r\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110869 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fdc18634-337d-4562-a6e5-a20df1f53028-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.110888 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdc18634-337d-4562-a6e5-a20df1f53028-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.217798 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdc18634-337d-4562-a6e5-a20df1f53028-config\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.217878 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.217922 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.217964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnm7r\" (UniqueName: \"kubernetes.io/projected/fdc18634-337d-4562-a6e5-a20df1f53028-kube-api-access-qnm7r\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.217993 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fdc18634-337d-4562-a6e5-a20df1f53028-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.218027 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdc18634-337d-4562-a6e5-a20df1f53028-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.218095 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.218156 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.220741 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdc18634-337d-4562-a6e5-a20df1f53028-config\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.220781 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fdc18634-337d-4562-a6e5-a20df1f53028-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.220856 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.221494 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdc18634-337d-4562-a6e5-a20df1f53028-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.230001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.236775 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.242516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdc18634-337d-4562-a6e5-a20df1f53028-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.243280 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnm7r\" (UniqueName: \"kubernetes.io/projected/fdc18634-337d-4562-a6e5-a20df1f53028-kube-api-access-qnm7r\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.258294 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fdc18634-337d-4562-a6e5-a20df1f53028\") " pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:37 crc kubenswrapper[4926]: I1007 21:11:37.342549 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.096293 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.096839 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.097000 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.102.83.12:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fq55l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-notifications-server-0_openstack(5b0d7388-05f4-4df6-b093-c9157a4658b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.098294 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-notifications-server-0" podUID="5b0d7388-05f4-4df6-b093-c9157a4658b0" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.123414 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.123481 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.123635 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.102.83.12:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5lls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(a9d51a12-f1e2-4b2e-a5d7-51474b31edb0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:42 crc kubenswrapper[4926]: E1007 21:11:42.125060 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" Oct 07 21:11:47 crc kubenswrapper[4926]: I1007 21:11:47.466869 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.774075 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.774135 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.774287 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4hr4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6d9454c9c9-z28fx_openstack(2c31b57b-e4b4-4a8f-945a-ef221de9dec4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.775487 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" podUID="2c31b57b-e4b4-4a8f-945a-ef221de9dec4" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.801303 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.801357 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.801468 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5c7h56dh5cfh8bh54fhbbhf4h5b9hdch67fhd7h55fh55fh6ch9h548h54ch665h647h6h8fhd6h5dfh5cdh58bh577h66fh695h5fbh55h77h5fcq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-76lls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-54b6445897-mtczn_openstack(76602401-809f-4d0f-8683-171a94389640): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.802654 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-54b6445897-mtczn" podUID="76602401-809f-4d0f-8683-171a94389640" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.808590 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.808614 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.808693 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zcv4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5c64b7c769-75lt2_openstack(4d47eec2-94cc-47d5-90fe-a3d17b72ec99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.809812 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" podUID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.853067 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.853260 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.853376 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mll45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5544966479-schqv_openstack(d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.854550 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5544966479-schqv" podUID="d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.891772 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.891824 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.891955 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ptbrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567d68c74f-df6kp_openstack(db1ff3fe-9ce5-455c-a414-a3c3a042bd39): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.894211 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" podUID="db1ff3fe-9ce5-455c-a414-a3c3a042bd39" Oct 07 21:11:47 crc kubenswrapper[4926]: I1007 21:11:47.986906 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fdc18634-337d-4562-a6e5-a20df1f53028","Type":"ContainerStarted","Data":"615235eecee144ef99cb30fc9c2dc00822f94cfe200e1cd8c6a8fd3fb458719b"} Oct 07 21:11:47 crc kubenswrapper[4926]: E1007 21:11:47.996800 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest\\\"\"" pod="openstack/dnsmasq-dns-54b6445897-mtczn" podUID="76602401-809f-4d0f-8683-171a94389640" Oct 07 21:11:48 crc kubenswrapper[4926]: E1007 21:11:47.999143 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/podified-master-centos10/openstack-neutron-server:watcher_latest\\\"\"" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" podUID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.716971 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 21:11:48 crc kubenswrapper[4926]: W1007 21:11:48.812744 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bf0bb92_3a4f_44ae_be03_e260cb246adb.slice/crio-c354f7c820e94a2f931873445d83c0fa445357a6e2cd9de7e8c3408c215af902 WatchSource:0}: Error finding container c354f7c820e94a2f931873445d83c0fa445357a6e2cd9de7e8c3408c215af902: Status 404 returned error can't find the container with id c354f7c820e94a2f931873445d83c0fa445357a6e2cd9de7e8c3408c215af902 Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.816033 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.841164 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.849017 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:11:48 crc kubenswrapper[4926]: W1007 21:11:48.856653 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ad789bb_88f9_4bec_ba17_46ac114893aa.slice/crio-5ffeef0a014c372039674a192f6a48d3a27f9f44a2b284f43f137b784b4c49df WatchSource:0}: Error finding container 5ffeef0a014c372039674a192f6a48d3a27f9f44a2b284f43f137b784b4c49df: Status 404 returned error can't find the container with id 5ffeef0a014c372039674a192f6a48d3a27f9f44a2b284f43f137b784b4c49df Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.867470 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:11:48 crc kubenswrapper[4926]: W1007 21:11:48.872090 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod038e2e94_c12a_4d54_8b48_96d08012ab97.slice/crio-cf9c8c4dea2921b7527b6404fb961310ea4361dad524082af3d71ff00b9aab7c WatchSource:0}: Error finding container cf9c8c4dea2921b7527b6404fb961310ea4361dad524082af3d71ff00b9aab7c: Status 404 returned error can't find the container with id cf9c8c4dea2921b7527b6404fb961310ea4361dad524082af3d71ff00b9aab7c Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.906714 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zg8nq"] Oct 07 21:11:48 crc kubenswrapper[4926]: W1007 21:11:48.907867 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod188a9695_679b_46c5_ac4b_26f71c9b4110.slice/crio-58143406c86f173cf3c7112295ef2a7f23b9d8b2763e18f59be1dc5763aa6a36 WatchSource:0}: Error finding container 58143406c86f173cf3c7112295ef2a7f23b9d8b2763e18f59be1dc5763aa6a36: Status 404 returned error can't find the container with id 58143406c86f173cf3c7112295ef2a7f23b9d8b2763e18f59be1dc5763aa6a36 Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.982396 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6qkqt"] Oct 07 21:11:48 crc kubenswrapper[4926]: I1007 21:11:48.990822 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.003488 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerStarted","Data":"5ffeef0a014c372039674a192f6a48d3a27f9f44a2b284f43f137b784b4c49df"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.005113 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" event={"ID":"2c31b57b-e4b4-4a8f-945a-ef221de9dec4","Type":"ContainerDied","Data":"11083f4b35787ab3e43d1fc97df3597515cf00327f5ec9b92c522b32c17dc4c1"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.005142 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11083f4b35787ab3e43d1fc97df3597515cf00327f5ec9b92c522b32c17dc4c1" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.007913 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"4c2c0ad449595a88cf92d0a873021ac44f258f8c38778755a129cc3c50f40b2f"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.010612 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" event={"ID":"db1ff3fe-9ce5-455c-a414-a3c3a042bd39","Type":"ContainerDied","Data":"1c4893cacf23d4c0b89c6dde170d81c477c025dc66f0854420568b399a56da7a"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.010638 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c4893cacf23d4c0b89c6dde170d81c477c025dc66f0854420568b399a56da7a" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.012207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b2704dcf-cd03-4f91-af4e-ea708f3c0534","Type":"ContainerStarted","Data":"bfc2b7bbd9d4869408558ea2ee24b8bb50de48f6f5e5869b1d55c05f0ecf283a"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.013597 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5544966479-schqv" event={"ID":"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc","Type":"ContainerDied","Data":"608448ab46333326d99c24420340b161d0fdea32f984711ed5b4fc885e71da5c"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.013639 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="608448ab46333326d99c24420340b161d0fdea32f984711ed5b4fc885e71da5c" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.014725 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7bf0bb92-3a4f-44ae-be03-e260cb246adb","Type":"ContainerStarted","Data":"c354f7c820e94a2f931873445d83c0fa445357a6e2cd9de7e8c3408c215af902"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.015881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"038e2e94-c12a-4d54-8b48-96d08012ab97","Type":"ContainerStarted","Data":"cf9c8c4dea2921b7527b6404fb961310ea4361dad524082af3d71ff00b9aab7c"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.017077 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zg8nq" event={"ID":"188a9695-679b-46c5-ac4b-26f71c9b4110","Type":"ContainerStarted","Data":"58143406c86f173cf3c7112295ef2a7f23b9d8b2763e18f59be1dc5763aa6a36"} Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.018226 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f838140a-347d-4bb3-8cc0-b4100bfe4b35","Type":"ContainerStarted","Data":"5e5fc466a100ccc586d1c0fa2af9116b839ec50d83e10c27163504e1e4786994"} Oct 07 21:11:49 crc kubenswrapper[4926]: W1007 21:11:49.090363 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d169a79_a57d_458f_a790_3623f6dd0885.slice/crio-7ba1618f8f06cc29ad222505fa8c1e98e96ce51c80b96c7ab9a45b76483023ff WatchSource:0}: Error finding container 7ba1618f8f06cc29ad222505fa8c1e98e96ce51c80b96c7ab9a45b76483023ff: Status 404 returned error can't find the container with id 7ba1618f8f06cc29ad222505fa8c1e98e96ce51c80b96c7ab9a45b76483023ff Oct 07 21:11:49 crc kubenswrapper[4926]: W1007 21:11:49.174823 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42ae43bc_ecb7_4cb8_95e5_394f7bc04dc2.slice/crio-1e1abe7d641ff70a8befa6d04ed1c37b6e10642eda876df1723964ab69e88609 WatchSource:0}: Error finding container 1e1abe7d641ff70a8befa6d04ed1c37b6e10642eda876df1723964ab69e88609: Status 404 returned error can't find the container with id 1e1abe7d641ff70a8befa6d04ed1c37b6e10642eda876df1723964ab69e88609 Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.275519 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.285363 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.285455 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.328533 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-config\") pod \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.329101 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-config" (OuterVolumeSpecName: "config") pod "2c31b57b-e4b4-4a8f-945a-ef221de9dec4" (UID: "2c31b57b-e4b4-4a8f-945a-ef221de9dec4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.329353 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hr4v\" (UniqueName: \"kubernetes.io/projected/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-kube-api-access-4hr4v\") pod \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\" (UID: \"2c31b57b-e4b4-4a8f-945a-ef221de9dec4\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.329382 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-dns-svc\") pod \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.330524 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptbrp\" (UniqueName: \"kubernetes.io/projected/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-kube-api-access-ptbrp\") pod \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.330570 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mll45\" (UniqueName: \"kubernetes.io/projected/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-kube-api-access-mll45\") pod \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.330599 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-dns-svc\") pod \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.330621 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-config\") pod \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\" (UID: \"d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.330643 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-config\") pod \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\" (UID: \"db1ff3fe-9ce5-455c-a414-a3c3a042bd39\") " Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.331076 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.329911 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db1ff3fe-9ce5-455c-a414-a3c3a042bd39" (UID: "db1ff3fe-9ce5-455c-a414-a3c3a042bd39"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.331461 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-config" (OuterVolumeSpecName: "config") pod "db1ff3fe-9ce5-455c-a414-a3c3a042bd39" (UID: "db1ff3fe-9ce5-455c-a414-a3c3a042bd39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.331856 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc" (UID: "d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.332212 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-config" (OuterVolumeSpecName: "config") pod "d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc" (UID: "d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.351449 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-kube-api-access-4hr4v" (OuterVolumeSpecName: "kube-api-access-4hr4v") pod "2c31b57b-e4b4-4a8f-945a-ef221de9dec4" (UID: "2c31b57b-e4b4-4a8f-945a-ef221de9dec4"). InnerVolumeSpecName "kube-api-access-4hr4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.358302 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-kube-api-access-mll45" (OuterVolumeSpecName: "kube-api-access-mll45") pod "d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc" (UID: "d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc"). InnerVolumeSpecName "kube-api-access-mll45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.362441 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-kube-api-access-ptbrp" (OuterVolumeSpecName: "kube-api-access-ptbrp") pod "db1ff3fe-9ce5-455c-a414-a3c3a042bd39" (UID: "db1ff3fe-9ce5-455c-a414-a3c3a042bd39"). InnerVolumeSpecName "kube-api-access-ptbrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.431948 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hr4v\" (UniqueName: \"kubernetes.io/projected/2c31b57b-e4b4-4a8f-945a-ef221de9dec4-kube-api-access-4hr4v\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.431986 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.431995 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptbrp\" (UniqueName: \"kubernetes.io/projected/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-kube-api-access-ptbrp\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.432014 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mll45\" (UniqueName: \"kubernetes.io/projected/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-kube-api-access-mll45\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.432027 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.432034 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:49 crc kubenswrapper[4926]: I1007 21:11:49.432042 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db1ff3fe-9ce5-455c-a414-a3c3a042bd39-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.026084 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0","Type":"ContainerStarted","Data":"e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad"} Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.030158 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"5b0d7388-05f4-4df6-b093-c9157a4658b0","Type":"ContainerStarted","Data":"c8c5d091362db91e9aa805fa9aaf324013dc391f8d44fa143c6039bb4fb9af52"} Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.031795 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6qkqt" event={"ID":"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2","Type":"ContainerStarted","Data":"1e1abe7d641ff70a8befa6d04ed1c37b6e10642eda876df1723964ab69e88609"} Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.033297 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57","Type":"ContainerStarted","Data":"aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231"} Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.034638 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567d68c74f-df6kp" Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.038300 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d169a79-a57d-458f-a790-3623f6dd0885","Type":"ContainerStarted","Data":"7ba1618f8f06cc29ad222505fa8c1e98e96ce51c80b96c7ab9a45b76483023ff"} Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.038405 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9454c9c9-z28fx" Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.038715 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5544966479-schqv" Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.137588 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9454c9c9-z28fx"] Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.145301 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d9454c9c9-z28fx"] Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.170458 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567d68c74f-df6kp"] Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.179280 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567d68c74f-df6kp"] Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.201019 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5544966479-schqv"] Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.205367 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5544966479-schqv"] Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.696179 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c31b57b-e4b4-4a8f-945a-ef221de9dec4" path="/var/lib/kubelet/pods/2c31b57b-e4b4-4a8f-945a-ef221de9dec4/volumes" Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.696771 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc" path="/var/lib/kubelet/pods/d12e9cf1-65dc-4c7c-a563-35a76b6fd1bc/volumes" Oct 07 21:11:50 crc kubenswrapper[4926]: I1007 21:11:50.697128 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db1ff3fe-9ce5-455c-a414-a3c3a042bd39" path="/var/lib/kubelet/pods/db1ff3fe-9ce5-455c-a414-a3c3a042bd39/volumes" Oct 07 21:11:55 crc kubenswrapper[4926]: I1007 21:11:55.111744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b2704dcf-cd03-4f91-af4e-ea708f3c0534","Type":"ContainerStarted","Data":"e0dc2d65452802c72eb66fd8a0222b86fa0a2ae4f1f96203455830b0c532fc35"} Oct 07 21:11:55 crc kubenswrapper[4926]: I1007 21:11:55.114707 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7bf0bb92-3a4f-44ae-be03-e260cb246adb","Type":"ContainerStarted","Data":"a638a71554890313c41804ba947ba293768ab709764aaec670046d0e1ff316de"} Oct 07 21:11:55 crc kubenswrapper[4926]: I1007 21:11:55.117184 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"038e2e94-c12a-4d54-8b48-96d08012ab97","Type":"ContainerStarted","Data":"7ca77056463cae76c9e66acca3925965ceb3a87e3d4c40b5a615ec8f11e0dbf8"} Oct 07 21:11:55 crc kubenswrapper[4926]: I1007 21:11:55.117385 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 07 21:11:55 crc kubenswrapper[4926]: I1007 21:11:55.159661 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.824988018 podStartE2EDuration="28.159642021s" podCreationTimestamp="2025-10-07 21:11:27 +0000 UTC" firstStartedPulling="2025-10-07 21:11:48.886553037 +0000 UTC m=+958.925134187" lastFinishedPulling="2025-10-07 21:11:54.22120704 +0000 UTC m=+964.259788190" observedRunningTime="2025-10-07 21:11:55.154515642 +0000 UTC m=+965.193096792" watchObservedRunningTime="2025-10-07 21:11:55.159642021 +0000 UTC m=+965.198223171" Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.126798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d169a79-a57d-458f-a790-3623f6dd0885","Type":"ContainerStarted","Data":"78e22f40976f7e8041418989401e1915b1678497bd327ce91ee1ec11aafa8c91"} Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.131323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fdc18634-337d-4562-a6e5-a20df1f53028","Type":"ContainerStarted","Data":"7ee3a4370b629a3d43cf0aed7d28b8c68c3a1578a6967d52124e544cb3f68865"} Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.136376 4926 generic.go:334] "Generic (PLEG): container finished" podID="188a9695-679b-46c5-ac4b-26f71c9b4110" containerID="6fc472fa21c76edfd57b28a121a2d97c29508b5e513d85d02cec0a759e15c02a" exitCode=0 Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.136454 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zg8nq" event={"ID":"188a9695-679b-46c5-ac4b-26f71c9b4110","Type":"ContainerDied","Data":"6fc472fa21c76edfd57b28a121a2d97c29508b5e513d85d02cec0a759e15c02a"} Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.141613 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6qkqt" event={"ID":"42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2","Type":"ContainerStarted","Data":"d409896f0b1564eea78bb13b5d5e9332f2eb83f5d1a5823371cdf4b6ec55bb27"} Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.142115 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6qkqt" Oct 07 21:11:56 crc kubenswrapper[4926]: I1007 21:11:56.168086 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6qkqt" podStartSLOduration=18.156702932 podStartE2EDuration="23.168070324s" podCreationTimestamp="2025-10-07 21:11:33 +0000 UTC" firstStartedPulling="2025-10-07 21:11:49.227701436 +0000 UTC m=+959.266282586" lastFinishedPulling="2025-10-07 21:11:54.239068788 +0000 UTC m=+964.277649978" observedRunningTime="2025-10-07 21:11:56.165300054 +0000 UTC m=+966.203881214" watchObservedRunningTime="2025-10-07 21:11:56.168070324 +0000 UTC m=+966.206651464" Oct 07 21:11:57 crc kubenswrapper[4926]: I1007 21:11:57.155270 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zg8nq" event={"ID":"188a9695-679b-46c5-ac4b-26f71c9b4110","Type":"ContainerStarted","Data":"ab163ee091c8ac55c74c08d4f9eeb7d59c49155742813548a1cba8ea969515ea"} Oct 07 21:11:57 crc kubenswrapper[4926]: I1007 21:11:57.156163 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zg8nq" event={"ID":"188a9695-679b-46c5-ac4b-26f71c9b4110","Type":"ContainerStarted","Data":"0ec760c85a8ac0741b17324e018578eb72db686990a30e73f9e06fb3cd5d4bc6"} Oct 07 21:11:57 crc kubenswrapper[4926]: I1007 21:11:57.156610 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:57 crc kubenswrapper[4926]: I1007 21:11:57.156646 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:11:57 crc kubenswrapper[4926]: I1007 21:11:57.177366 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-zg8nq" podStartSLOduration=19.035452953 podStartE2EDuration="24.177341842s" podCreationTimestamp="2025-10-07 21:11:33 +0000 UTC" firstStartedPulling="2025-10-07 21:11:48.911679976 +0000 UTC m=+958.950261126" lastFinishedPulling="2025-10-07 21:11:54.053568845 +0000 UTC m=+964.092150015" observedRunningTime="2025-10-07 21:11:57.173419628 +0000 UTC m=+967.212000798" watchObservedRunningTime="2025-10-07 21:11:57.177341842 +0000 UTC m=+967.215922992" Oct 07 21:11:58 crc kubenswrapper[4926]: I1007 21:11:58.166919 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerStarted","Data":"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08"} Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.191868 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f838140a-347d-4bb3-8cc0-b4100bfe4b35","Type":"ContainerStarted","Data":"c5849ec77af20d02832aac47aa1f49f50eb8926ee7f12ca8b44110c99e92132d"} Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.194082 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.194886 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fdc18634-337d-4562-a6e5-a20df1f53028","Type":"ContainerStarted","Data":"adc3f00f09104acecb60208e2f76866e2b43abe145d9fa2ca12658e5c4b6840b"} Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.198232 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b2704dcf-cd03-4f91-af4e-ea708f3c0534","Type":"ContainerStarted","Data":"9a4fa66617b9daa8944e46517d1d7bfd6820fc6097934554abff70de0f317aeb"} Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.219435 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=20.792246729 podStartE2EDuration="31.219410098s" podCreationTimestamp="2025-10-07 21:11:29 +0000 UTC" firstStartedPulling="2025-10-07 21:11:48.886429263 +0000 UTC m=+958.925010413" lastFinishedPulling="2025-10-07 21:11:59.313592632 +0000 UTC m=+969.352173782" observedRunningTime="2025-10-07 21:12:00.21569278 +0000 UTC m=+970.254273940" watchObservedRunningTime="2025-10-07 21:12:00.219410098 +0000 UTC m=+970.257991268" Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.294602 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.690395874 podStartE2EDuration="25.294583509s" podCreationTimestamp="2025-10-07 21:11:35 +0000 UTC" firstStartedPulling="2025-10-07 21:11:47.80736355 +0000 UTC m=+957.845944700" lastFinishedPulling="2025-10-07 21:11:59.411551185 +0000 UTC m=+969.450132335" observedRunningTime="2025-10-07 21:12:00.262122357 +0000 UTC m=+970.300703557" watchObservedRunningTime="2025-10-07 21:12:00.294583509 +0000 UTC m=+970.333164659" Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.299740 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.744964306 podStartE2EDuration="28.299726948s" podCreationTimestamp="2025-10-07 21:11:32 +0000 UTC" firstStartedPulling="2025-10-07 21:11:48.842935411 +0000 UTC m=+958.881516561" lastFinishedPulling="2025-10-07 21:11:59.397698043 +0000 UTC m=+969.436279203" observedRunningTime="2025-10-07 21:12:00.292126448 +0000 UTC m=+970.330707598" watchObservedRunningTime="2025-10-07 21:12:00.299726948 +0000 UTC m=+970.338308098" Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.607341 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 07 21:12:00 crc kubenswrapper[4926]: I1007 21:12:00.656503 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.212559 4926 generic.go:334] "Generic (PLEG): container finished" podID="7bf0bb92-3a4f-44ae-be03-e260cb246adb" containerID="a638a71554890313c41804ba947ba293768ab709764aaec670046d0e1ff316de" exitCode=0 Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.212703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7bf0bb92-3a4f-44ae-be03-e260cb246adb","Type":"ContainerDied","Data":"a638a71554890313c41804ba947ba293768ab709764aaec670046d0e1ff316de"} Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.213276 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.267531 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.343037 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.411148 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.540518 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b6445897-mtczn"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.585933 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bd9dbd8c9-xjt25"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.594282 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.597687 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd9dbd8c9-xjt25"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.598800 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.664303 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-5g8lv"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.665645 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.667551 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.681785 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5g8lv"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.684378 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.684511 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-dns-svc\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.684680 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkc45\" (UniqueName: \"kubernetes.io/projected/79580d2e-279c-4014-9b7e-da8da257d48c-kube-api-access-nkc45\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.684792 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-config\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791450 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5r2\" (UniqueName: \"kubernetes.io/projected/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-kube-api-access-5n5r2\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791504 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791573 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-dns-svc\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791635 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-config\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791920 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-combined-ca-bundle\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkc45\" (UniqueName: \"kubernetes.io/projected/79580d2e-279c-4014-9b7e-da8da257d48c-kube-api-access-nkc45\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.791959 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-config\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.792001 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-ovs-rundir\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.792025 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-ovn-rundir\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.792657 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.792940 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-dns-svc\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.793534 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-config\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.809132 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkc45\" (UniqueName: \"kubernetes.io/projected/79580d2e-279c-4014-9b7e-da8da257d48c-kube-api-access-nkc45\") pod \"dnsmasq-dns-5bd9dbd8c9-xjt25\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.847589 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c64b7c769-75lt2"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.858259 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c9b5c568f-t9jpg"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.859846 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.863366 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.878475 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9b5c568f-t9jpg"] Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893153 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893221 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893251 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-config\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893282 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-combined-ca-bundle\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893325 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-config\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893348 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-ovs-rundir\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893366 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-ovn-rundir\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893409 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-dns-svc\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893447 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5r2\" (UniqueName: \"kubernetes.io/projected/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-kube-api-access-5n5r2\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.893476 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6tc8\" (UniqueName: \"kubernetes.io/projected/869188a9-80d3-454b-8e54-0e885881463f-kube-api-access-f6tc8\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.896387 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-ovs-rundir\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.896446 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-ovn-rundir\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.896747 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-config\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.899358 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.900934 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-combined-ca-bundle\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.913376 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5r2\" (UniqueName: \"kubernetes.io/projected/5a9488ef-d520-4ff8-9eb2-8de1c72f03ce-kube-api-access-5n5r2\") pod \"ovn-controller-metrics-5g8lv\" (UID: \"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce\") " pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.936792 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.943876 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.987920 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5g8lv" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994156 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-dns-svc\") pod \"76602401-809f-4d0f-8683-171a94389640\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994336 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-config\") pod \"76602401-809f-4d0f-8683-171a94389640\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994404 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76lls\" (UniqueName: \"kubernetes.io/projected/76602401-809f-4d0f-8683-171a94389640-kube-api-access-76lls\") pod \"76602401-809f-4d0f-8683-171a94389640\" (UID: \"76602401-809f-4d0f-8683-171a94389640\") " Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994608 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-config\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994642 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994667 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-dns-svc\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994718 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6tc8\" (UniqueName: \"kubernetes.io/projected/869188a9-80d3-454b-8e54-0e885881463f-kube-api-access-f6tc8\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.994751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.995936 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.996302 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-config" (OuterVolumeSpecName: "config") pod "76602401-809f-4d0f-8683-171a94389640" (UID: "76602401-809f-4d0f-8683-171a94389640"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.996782 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-config\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.996608 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "76602401-809f-4d0f-8683-171a94389640" (UID: "76602401-809f-4d0f-8683-171a94389640"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.996909 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-dns-svc\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:01 crc kubenswrapper[4926]: I1007 21:12:01.997939 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.000651 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76602401-809f-4d0f-8683-171a94389640-kube-api-access-76lls" (OuterVolumeSpecName: "kube-api-access-76lls") pod "76602401-809f-4d0f-8683-171a94389640" (UID: "76602401-809f-4d0f-8683-171a94389640"). InnerVolumeSpecName "kube-api-access-76lls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.013802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6tc8\" (UniqueName: \"kubernetes.io/projected/869188a9-80d3-454b-8e54-0e885881463f-kube-api-access-f6tc8\") pod \"dnsmasq-dns-c9b5c568f-t9jpg\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.096361 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76lls\" (UniqueName: \"kubernetes.io/projected/76602401-809f-4d0f-8683-171a94389640-kube-api-access-76lls\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.096743 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.096756 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76602401-809f-4d0f-8683-171a94389640-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.201997 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.232027 4926 generic.go:334] "Generic (PLEG): container finished" podID="2d169a79-a57d-458f-a790-3623f6dd0885" containerID="78e22f40976f7e8041418989401e1915b1678497bd327ce91ee1ec11aafa8c91" exitCode=0 Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.232104 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d169a79-a57d-458f-a790-3623f6dd0885","Type":"ContainerDied","Data":"78e22f40976f7e8041418989401e1915b1678497bd327ce91ee1ec11aafa8c91"} Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.250860 4926 generic.go:334] "Generic (PLEG): container finished" podID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" containerID="9604ec0d7f00046aa3a89fbe67b85574aff84224666fdc36eba8da3289c5aa75" exitCode=0 Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.250975 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" event={"ID":"4d47eec2-94cc-47d5-90fe-a3d17b72ec99","Type":"ContainerDied","Data":"9604ec0d7f00046aa3a89fbe67b85574aff84224666fdc36eba8da3289c5aa75"} Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.254297 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7bf0bb92-3a4f-44ae-be03-e260cb246adb","Type":"ContainerStarted","Data":"df058a1bd6b596a0f802449f5ae92f16de5696a03e32df3fa57336b7bbc2982f"} Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.256906 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b6445897-mtczn" event={"ID":"76602401-809f-4d0f-8683-171a94389640","Type":"ContainerDied","Data":"472a4b10f82dcf3e3b9b078d94ff8b6e5134a2889eb9f690ad0aa921010e4aa1"} Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.257030 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b6445897-mtczn" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.258538 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.353372 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=32.114603691 podStartE2EDuration="37.353353871s" podCreationTimestamp="2025-10-07 21:11:25 +0000 UTC" firstStartedPulling="2025-10-07 21:11:48.814816945 +0000 UTC m=+958.853398085" lastFinishedPulling="2025-10-07 21:11:54.053567115 +0000 UTC m=+964.092148265" observedRunningTime="2025-10-07 21:12:02.326212483 +0000 UTC m=+972.364793633" watchObservedRunningTime="2025-10-07 21:12:02.353353871 +0000 UTC m=+972.391935031" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.357984 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 07 21:12:02 crc kubenswrapper[4926]: W1007 21:12:02.423501 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79580d2e_279c_4014_9b7e_da8da257d48c.slice/crio-e89414833bf0b8fea0707c4beafccc1909a0acb5b98501925e7e1887039a3584 WatchSource:0}: Error finding container e89414833bf0b8fea0707c4beafccc1909a0acb5b98501925e7e1887039a3584: Status 404 returned error can't find the container with id e89414833bf0b8fea0707c4beafccc1909a0acb5b98501925e7e1887039a3584 Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.429038 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd9dbd8c9-xjt25"] Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.446104 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b6445897-mtczn"] Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.452044 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54b6445897-mtczn"] Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.552133 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5g8lv"] Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.572224 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.574602 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.577814 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.577943 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-sqfxj" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.578079 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.578314 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.591473 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.600047 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.636964 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.698468 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76602401-809f-4d0f-8683-171a94389640" path="/var/lib/kubelet/pods/76602401-809f-4d0f-8683-171a94389640/volumes" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.712895 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-dns-svc\") pod \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.713256 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcv4f\" (UniqueName: \"kubernetes.io/projected/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-kube-api-access-zcv4f\") pod \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.713446 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-config\") pod \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\" (UID: \"4d47eec2-94cc-47d5-90fe-a3d17b72ec99\") " Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.713811 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee1a017c-a9f5-497d-9b4d-80f6f776be50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.713950 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1a017c-a9f5-497d-9b4d-80f6f776be50-config\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.714077 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.714183 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee1a017c-a9f5-497d-9b4d-80f6f776be50-scripts\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.714342 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-885fw\" (UniqueName: \"kubernetes.io/projected/ee1a017c-a9f5-497d-9b4d-80f6f776be50-kube-api-access-885fw\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.714456 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.714580 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.719935 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-kube-api-access-zcv4f" (OuterVolumeSpecName: "kube-api-access-zcv4f") pod "4d47eec2-94cc-47d5-90fe-a3d17b72ec99" (UID: "4d47eec2-94cc-47d5-90fe-a3d17b72ec99"). InnerVolumeSpecName "kube-api-access-zcv4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.733699 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d47eec2-94cc-47d5-90fe-a3d17b72ec99" (UID: "4d47eec2-94cc-47d5-90fe-a3d17b72ec99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.739325 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-config" (OuterVolumeSpecName: "config") pod "4d47eec2-94cc-47d5-90fe-a3d17b72ec99" (UID: "4d47eec2-94cc-47d5-90fe-a3d17b72ec99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816676 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee1a017c-a9f5-497d-9b4d-80f6f776be50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816730 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1a017c-a9f5-497d-9b4d-80f6f776be50-config\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816788 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816813 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee1a017c-a9f5-497d-9b4d-80f6f776be50-scripts\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-885fw\" (UniqueName: \"kubernetes.io/projected/ee1a017c-a9f5-497d-9b4d-80f6f776be50-kube-api-access-885fw\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.816905 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.817033 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.817045 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcv4f\" (UniqueName: \"kubernetes.io/projected/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-kube-api-access-zcv4f\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.817054 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d47eec2-94cc-47d5-90fe-a3d17b72ec99-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.818230 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee1a017c-a9f5-497d-9b4d-80f6f776be50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.820070 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee1a017c-a9f5-497d-9b4d-80f6f776be50-scripts\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.820487 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1a017c-a9f5-497d-9b4d-80f6f776be50-config\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.822897 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.823655 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.825557 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee1a017c-a9f5-497d-9b4d-80f6f776be50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.835862 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-885fw\" (UniqueName: \"kubernetes.io/projected/ee1a017c-a9f5-497d-9b4d-80f6f776be50-kube-api-access-885fw\") pod \"ovn-northd-0\" (UID: \"ee1a017c-a9f5-497d-9b4d-80f6f776be50\") " pod="openstack/ovn-northd-0" Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.856813 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9b5c568f-t9jpg"] Oct 07 21:12:02 crc kubenswrapper[4926]: W1007 21:12:02.863935 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod869188a9_80d3_454b_8e54_0e885881463f.slice/crio-413c4ed3a617bb5621dc9bb21ef80a2bcc31f7f3970cd0216df1404ac2f0c7fa WatchSource:0}: Error finding container 413c4ed3a617bb5621dc9bb21ef80a2bcc31f7f3970cd0216df1404ac2f0c7fa: Status 404 returned error can't find the container with id 413c4ed3a617bb5621dc9bb21ef80a2bcc31f7f3970cd0216df1404ac2f0c7fa Oct 07 21:12:02 crc kubenswrapper[4926]: I1007 21:12:02.900051 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.271553 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5g8lv" event={"ID":"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce","Type":"ContainerStarted","Data":"4d3aaaf2e8f8a1abbd5d6df7533b781f940bb9f0605acbf8422552fba5fa4711"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.271888 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5g8lv" event={"ID":"5a9488ef-d520-4ff8-9eb2-8de1c72f03ce","Type":"ContainerStarted","Data":"b9dc8666d07227cf418dd0169cc36deea1b8a6be7fd271e195e8a03f8bf6cecc"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.276012 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d169a79-a57d-458f-a790-3623f6dd0885","Type":"ContainerStarted","Data":"43a7f9463435751d007ac6a7c7124dbd2679c7d10becbd95ed21b1e9fba7a80b"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.277426 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.277425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64b7c769-75lt2" event={"ID":"4d47eec2-94cc-47d5-90fe-a3d17b72ec99","Type":"ContainerDied","Data":"4aa4674218b8e29e9d4dd2eead33f7b29f9ff6f2b0bd67590cb8466a49de642c"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.277537 4926 scope.go:117] "RemoveContainer" containerID="9604ec0d7f00046aa3a89fbe67b85574aff84224666fdc36eba8da3289c5aa75" Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.279158 4926 generic.go:334] "Generic (PLEG): container finished" podID="869188a9-80d3-454b-8e54-0e885881463f" containerID="351555bd14f81eb8b0f93340e06e4d042de73cc132d3495b66c58f0be22f80ab" exitCode=0 Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.279208 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" event={"ID":"869188a9-80d3-454b-8e54-0e885881463f","Type":"ContainerDied","Data":"351555bd14f81eb8b0f93340e06e4d042de73cc132d3495b66c58f0be22f80ab"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.279243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" event={"ID":"869188a9-80d3-454b-8e54-0e885881463f","Type":"ContainerStarted","Data":"413c4ed3a617bb5621dc9bb21ef80a2bcc31f7f3970cd0216df1404ac2f0c7fa"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.284872 4926 generic.go:334] "Generic (PLEG): container finished" podID="79580d2e-279c-4014-9b7e-da8da257d48c" containerID="a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb" exitCode=0 Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.286258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" event={"ID":"79580d2e-279c-4014-9b7e-da8da257d48c","Type":"ContainerDied","Data":"a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.286317 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" event={"ID":"79580d2e-279c-4014-9b7e-da8da257d48c","Type":"ContainerStarted","Data":"e89414833bf0b8fea0707c4beafccc1909a0acb5b98501925e7e1887039a3584"} Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.308073 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-5g8lv" podStartSLOduration=2.308053615 podStartE2EDuration="2.308053615s" podCreationTimestamp="2025-10-07 21:12:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:03.298460377 +0000 UTC m=+973.337041527" watchObservedRunningTime="2025-10-07 21:12:03.308053615 +0000 UTC m=+973.346634765" Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.356556 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.372356 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=32.931163526 podStartE2EDuration="38.37233263s" podCreationTimestamp="2025-10-07 21:11:25 +0000 UTC" firstStartedPulling="2025-10-07 21:11:49.092802552 +0000 UTC m=+959.131383702" lastFinishedPulling="2025-10-07 21:11:54.533971656 +0000 UTC m=+964.572552806" observedRunningTime="2025-10-07 21:12:03.352859905 +0000 UTC m=+973.391441055" watchObservedRunningTime="2025-10-07 21:12:03.37233263 +0000 UTC m=+973.410913780" Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.510179 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c64b7c769-75lt2"] Oct 07 21:12:03 crc kubenswrapper[4926]: I1007 21:12:03.511831 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c64b7c769-75lt2"] Oct 07 21:12:04 crc kubenswrapper[4926]: I1007 21:12:04.301736 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ee1a017c-a9f5-497d-9b4d-80f6f776be50","Type":"ContainerStarted","Data":"c15f800bccbcfe001bb8e1836a0f5c965658a6ebad4b75530bc9f23b8a3ff23f"} Oct 07 21:12:04 crc kubenswrapper[4926]: I1007 21:12:04.690053 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" path="/var/lib/kubelet/pods/4d47eec2-94cc-47d5-90fe-a3d17b72ec99/volumes" Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.320528 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" event={"ID":"869188a9-80d3-454b-8e54-0e885881463f","Type":"ContainerStarted","Data":"e4bdac8a8da435e799222ab498f246d4a44a31aa41525ed182770d2dfb8c79e5"} Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.321239 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.323985 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" event={"ID":"79580d2e-279c-4014-9b7e-da8da257d48c","Type":"ContainerStarted","Data":"f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833"} Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.324102 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.328213 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ee1a017c-a9f5-497d-9b4d-80f6f776be50","Type":"ContainerStarted","Data":"15b4217a026b044b8e5d1d32837315a3a99892d0efbed556df39477d5dfe8ae9"} Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.328256 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ee1a017c-a9f5-497d-9b4d-80f6f776be50","Type":"ContainerStarted","Data":"48c8162049c645679bc535c1609ff7f368fac8d382dda861910ef5649dd39252"} Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.328369 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.352911 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" podStartSLOduration=4.352891293 podStartE2EDuration="4.352891293s" podCreationTimestamp="2025-10-07 21:12:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:05.346466967 +0000 UTC m=+975.385048177" watchObservedRunningTime="2025-10-07 21:12:05.352891293 +0000 UTC m=+975.391472473" Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.379128 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" podStartSLOduration=4.379107524 podStartE2EDuration="4.379107524s" podCreationTimestamp="2025-10-07 21:12:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:05.366246431 +0000 UTC m=+975.404827661" watchObservedRunningTime="2025-10-07 21:12:05.379107524 +0000 UTC m=+975.417688684" Oct 07 21:12:05 crc kubenswrapper[4926]: I1007 21:12:05.407800 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.02777345 podStartE2EDuration="3.407773306s" podCreationTimestamp="2025-10-07 21:12:02 +0000 UTC" firstStartedPulling="2025-10-07 21:12:03.387912952 +0000 UTC m=+973.426494102" lastFinishedPulling="2025-10-07 21:12:04.767912798 +0000 UTC m=+974.806493958" observedRunningTime="2025-10-07 21:12:05.400451824 +0000 UTC m=+975.439033054" watchObservedRunningTime="2025-10-07 21:12:05.407773306 +0000 UTC m=+975.446354496" Oct 07 21:12:06 crc kubenswrapper[4926]: I1007 21:12:06.340232 4926 generic.go:334] "Generic (PLEG): container finished" podID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerID="572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08" exitCode=0 Oct 07 21:12:06 crc kubenswrapper[4926]: I1007 21:12:06.340357 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerDied","Data":"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08"} Oct 07 21:12:07 crc kubenswrapper[4926]: I1007 21:12:07.458763 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 07 21:12:07 crc kubenswrapper[4926]: I1007 21:12:07.459059 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 07 21:12:07 crc kubenswrapper[4926]: I1007 21:12:07.510747 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 07 21:12:07 crc kubenswrapper[4926]: I1007 21:12:07.510800 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 07 21:12:07 crc kubenswrapper[4926]: I1007 21:12:07.747157 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 07 21:12:08 crc kubenswrapper[4926]: I1007 21:12:08.433560 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.478726 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.572153 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd9dbd8c9-xjt25"] Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.572365 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" containerName="dnsmasq-dns" containerID="cri-o://f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833" gracePeriod=10 Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.575557 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.599116 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67cfd59845-5psqc"] Oct 07 21:12:09 crc kubenswrapper[4926]: E1007 21:12:09.599498 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" containerName="init" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.599513 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" containerName="init" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.599697 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d47eec2-94cc-47d5-90fe-a3d17b72ec99" containerName="init" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.600551 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.704264 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67cfd59845-5psqc"] Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.729117 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-4rl9j"] Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.730355 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.737711 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-4rl9j"] Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.756355 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-dns-svc\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.756434 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-sb\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.756478 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-config\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.756536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lpjj\" (UniqueName: \"kubernetes.io/projected/8854b1d1-312d-42cf-b908-182769253998-kube-api-access-8lpjj\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.756553 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-nb\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.809143 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.858132 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-config\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.858248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lpjj\" (UniqueName: \"kubernetes.io/projected/8854b1d1-312d-42cf-b908-182769253998-kube-api-access-8lpjj\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.858280 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-nb\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.858340 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tccgh\" (UniqueName: \"kubernetes.io/projected/51cda538-9dad-4aef-91ac-16af8c38e2dd-kube-api-access-tccgh\") pod \"watcher-db-create-4rl9j\" (UID: \"51cda538-9dad-4aef-91ac-16af8c38e2dd\") " pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.858393 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-dns-svc\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.858442 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-sb\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.860392 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-config\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.860965 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-dns-svc\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.861223 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-nb\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.866312 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-sb\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.882733 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lpjj\" (UniqueName: \"kubernetes.io/projected/8854b1d1-312d-42cf-b908-182769253998-kube-api-access-8lpjj\") pod \"dnsmasq-dns-67cfd59845-5psqc\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.904537 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.956097 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.961072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tccgh\" (UniqueName: \"kubernetes.io/projected/51cda538-9dad-4aef-91ac-16af8c38e2dd-kube-api-access-tccgh\") pod \"watcher-db-create-4rl9j\" (UID: \"51cda538-9dad-4aef-91ac-16af8c38e2dd\") " pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:09 crc kubenswrapper[4926]: I1007 21:12:09.988527 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tccgh\" (UniqueName: \"kubernetes.io/projected/51cda538-9dad-4aef-91ac-16af8c38e2dd-kube-api-access-tccgh\") pod \"watcher-db-create-4rl9j\" (UID: \"51cda538-9dad-4aef-91ac-16af8c38e2dd\") " pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.049638 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.101531 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.265646 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkc45\" (UniqueName: \"kubernetes.io/projected/79580d2e-279c-4014-9b7e-da8da257d48c-kube-api-access-nkc45\") pod \"79580d2e-279c-4014-9b7e-da8da257d48c\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.265724 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-config\") pod \"79580d2e-279c-4014-9b7e-da8da257d48c\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.265796 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-dns-svc\") pod \"79580d2e-279c-4014-9b7e-da8da257d48c\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.265888 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-ovsdbserver-nb\") pod \"79580d2e-279c-4014-9b7e-da8da257d48c\" (UID: \"79580d2e-279c-4014-9b7e-da8da257d48c\") " Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.271390 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79580d2e-279c-4014-9b7e-da8da257d48c-kube-api-access-nkc45" (OuterVolumeSpecName: "kube-api-access-nkc45") pod "79580d2e-279c-4014-9b7e-da8da257d48c" (UID: "79580d2e-279c-4014-9b7e-da8da257d48c"). InnerVolumeSpecName "kube-api-access-nkc45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.307686 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79580d2e-279c-4014-9b7e-da8da257d48c" (UID: "79580d2e-279c-4014-9b7e-da8da257d48c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.307959 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-config" (OuterVolumeSpecName: "config") pod "79580d2e-279c-4014-9b7e-da8da257d48c" (UID: "79580d2e-279c-4014-9b7e-da8da257d48c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.316613 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "79580d2e-279c-4014-9b7e-da8da257d48c" (UID: "79580d2e-279c-4014-9b7e-da8da257d48c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.371447 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.371496 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.371511 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkc45\" (UniqueName: \"kubernetes.io/projected/79580d2e-279c-4014-9b7e-da8da257d48c-kube-api-access-nkc45\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.371524 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79580d2e-279c-4014-9b7e-da8da257d48c-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.389747 4926 generic.go:334] "Generic (PLEG): container finished" podID="79580d2e-279c-4014-9b7e-da8da257d48c" containerID="f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833" exitCode=0 Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.389844 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.389858 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" event={"ID":"79580d2e-279c-4014-9b7e-da8da257d48c","Type":"ContainerDied","Data":"f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833"} Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.389916 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd9dbd8c9-xjt25" event={"ID":"79580d2e-279c-4014-9b7e-da8da257d48c","Type":"ContainerDied","Data":"e89414833bf0b8fea0707c4beafccc1909a0acb5b98501925e7e1887039a3584"} Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.390654 4926 scope.go:117] "RemoveContainer" containerID="f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833" Oct 07 21:12:10 crc kubenswrapper[4926]: W1007 21:12:10.408490 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8854b1d1_312d_42cf_b908_182769253998.slice/crio-c86f6e125d65e6217d5e71ae91b297645fbc0242999fd07614802c58a6da0c64 WatchSource:0}: Error finding container c86f6e125d65e6217d5e71ae91b297645fbc0242999fd07614802c58a6da0c64: Status 404 returned error can't find the container with id c86f6e125d65e6217d5e71ae91b297645fbc0242999fd07614802c58a6da0c64 Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.412907 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67cfd59845-5psqc"] Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.429888 4926 scope.go:117] "RemoveContainer" containerID="a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.434608 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd9dbd8c9-xjt25"] Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.445019 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bd9dbd8c9-xjt25"] Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.474404 4926 scope.go:117] "RemoveContainer" containerID="f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833" Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.475481 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833\": container with ID starting with f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833 not found: ID does not exist" containerID="f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.475519 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833"} err="failed to get container status \"f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833\": rpc error: code = NotFound desc = could not find container \"f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833\": container with ID starting with f56435185775436e9884f5534bb400f0ba5f3e093371254befbab634dbc76833 not found: ID does not exist" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.475537 4926 scope.go:117] "RemoveContainer" containerID="a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb" Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.476895 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb\": container with ID starting with a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb not found: ID does not exist" containerID="a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.477015 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb"} err="failed to get container status \"a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb\": rpc error: code = NotFound desc = could not find container \"a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb\": container with ID starting with a2f0126674fe3540649dc1dd20a5c561009a0ea667483f0c5c173071b97ce7cb not found: ID does not exist" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.498665 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-4rl9j"] Oct 07 21:12:10 crc kubenswrapper[4926]: W1007 21:12:10.507811 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51cda538_9dad_4aef_91ac_16af8c38e2dd.slice/crio-1c8f4eb909a86a290a4e91bd2b2834976fec88d50a01d590abbfac4b1f7ddb59 WatchSource:0}: Error finding container 1c8f4eb909a86a290a4e91bd2b2834976fec88d50a01d590abbfac4b1f7ddb59: Status 404 returned error can't find the container with id 1c8f4eb909a86a290a4e91bd2b2834976fec88d50a01d590abbfac4b1f7ddb59 Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.694342 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" path="/var/lib/kubelet/pods/79580d2e-279c-4014-9b7e-da8da257d48c/volumes" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.703295 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.703610 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" containerName="init" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.703626 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" containerName="init" Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.703648 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" containerName="dnsmasq-dns" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.703655 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" containerName="dnsmasq-dns" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.703909 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="79580d2e-279c-4014-9b7e-da8da257d48c" containerName="dnsmasq-dns" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.709945 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.712970 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-2vz8x" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.713868 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.714030 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.714111 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.727608 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.889911 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.889974 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzzv8\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-kube-api-access-zzzv8\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.890036 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8671e7ac-5c56-4b89-9d72-386b75d12ec1-cache\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.890076 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8671e7ac-5c56-4b89-9d72-386b75d12ec1-lock\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.890131 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.991908 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8671e7ac-5c56-4b89-9d72-386b75d12ec1-cache\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.991990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8671e7ac-5c56-4b89-9d72-386b75d12ec1-lock\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.992038 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.992140 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.992167 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzzv8\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-kube-api-access-zzzv8\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.992355 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8671e7ac-5c56-4b89-9d72-386b75d12ec1-cache\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.992627 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.992627 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.992684 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 21:12:10 crc kubenswrapper[4926]: E1007 21:12:10.992754 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift podName:8671e7ac-5c56-4b89-9d72-386b75d12ec1 nodeName:}" failed. No retries permitted until 2025-10-07 21:12:11.492736672 +0000 UTC m=+981.531317822 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift") pod "swift-storage-0" (UID: "8671e7ac-5c56-4b89-9d72-386b75d12ec1") : configmap "swift-ring-files" not found Oct 07 21:12:10 crc kubenswrapper[4926]: I1007 21:12:10.992972 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8671e7ac-5c56-4b89-9d72-386b75d12ec1-lock\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.009604 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzzv8\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-kube-api-access-zzzv8\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.022577 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.173825 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-fz74j"] Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.174878 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.183390 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.183595 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.183733 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.230815 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fz74j"] Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297027 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw82m\" (UniqueName: \"kubernetes.io/projected/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-kube-api-access-qw82m\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297072 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-etc-swift\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297146 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-scripts\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297293 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-swiftconf\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297325 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-ring-data-devices\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297381 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-dispersionconf\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.297403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-combined-ca-bundle\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.398686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-swiftconf\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.398769 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-ring-data-devices\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.398869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-dispersionconf\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.398904 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-combined-ca-bundle\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.398945 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw82m\" (UniqueName: \"kubernetes.io/projected/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-kube-api-access-qw82m\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.398977 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-etc-swift\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.399028 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-scripts\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.399674 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-ring-data-devices\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.400022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-etc-swift\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.400112 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-scripts\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.401057 4926 generic.go:334] "Generic (PLEG): container finished" podID="51cda538-9dad-4aef-91ac-16af8c38e2dd" containerID="eed6a6831b08b5dd4f5319631afb91ba0ca92383e5039b7b6f0969dcb249b901" exitCode=0 Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.401287 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-4rl9j" event={"ID":"51cda538-9dad-4aef-91ac-16af8c38e2dd","Type":"ContainerDied","Data":"eed6a6831b08b5dd4f5319631afb91ba0ca92383e5039b7b6f0969dcb249b901"} Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.401395 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-4rl9j" event={"ID":"51cda538-9dad-4aef-91ac-16af8c38e2dd","Type":"ContainerStarted","Data":"1c8f4eb909a86a290a4e91bd2b2834976fec88d50a01d590abbfac4b1f7ddb59"} Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.402843 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-combined-ca-bundle\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.404670 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-dispersionconf\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.405409 4926 generic.go:334] "Generic (PLEG): container finished" podID="8854b1d1-312d-42cf-b908-182769253998" containerID="e4c837174b07a268ad0e0b01ada50cabcfa53b9a6244e4337b0120168f080d64" exitCode=0 Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.405462 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" event={"ID":"8854b1d1-312d-42cf-b908-182769253998","Type":"ContainerDied","Data":"e4c837174b07a268ad0e0b01ada50cabcfa53b9a6244e4337b0120168f080d64"} Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.405493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" event={"ID":"8854b1d1-312d-42cf-b908-182769253998","Type":"ContainerStarted","Data":"c86f6e125d65e6217d5e71ae91b297645fbc0242999fd07614802c58a6da0c64"} Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.416294 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-swiftconf\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.417784 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw82m\" (UniqueName: \"kubernetes.io/projected/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-kube-api-access-qw82m\") pod \"swift-ring-rebalance-fz74j\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.498696 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.500359 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:11 crc kubenswrapper[4926]: E1007 21:12:11.500584 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 21:12:11 crc kubenswrapper[4926]: E1007 21:12:11.500611 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 21:12:11 crc kubenswrapper[4926]: E1007 21:12:11.500691 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift podName:8671e7ac-5c56-4b89-9d72-386b75d12ec1 nodeName:}" failed. No retries permitted until 2025-10-07 21:12:12.50064428 +0000 UTC m=+982.539225430 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift") pod "swift-storage-0" (UID: "8671e7ac-5c56-4b89-9d72-386b75d12ec1") : configmap "swift-ring-files" not found Oct 07 21:12:11 crc kubenswrapper[4926]: I1007 21:12:11.967048 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fz74j"] Oct 07 21:12:11 crc kubenswrapper[4926]: W1007 21:12:11.972645 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf314e4d4_9c99_4880_ab6e_a2eb04e2ae5f.slice/crio-9b3d19b287c918b653f292794b4b016d0f6e1ab796aa94d841ad230b6bd396a1 WatchSource:0}: Error finding container 9b3d19b287c918b653f292794b4b016d0f6e1ab796aa94d841ad230b6bd396a1: Status 404 returned error can't find the container with id 9b3d19b287c918b653f292794b4b016d0f6e1ab796aa94d841ad230b6bd396a1 Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.204435 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.414475 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" event={"ID":"8854b1d1-312d-42cf-b908-182769253998","Type":"ContainerStarted","Data":"636d612a3231de1ed8a767cb9366b49da87eec095b11e9308ebbe9479aa6e1f0"} Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.414576 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.416137 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fz74j" event={"ID":"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f","Type":"ContainerStarted","Data":"9b3d19b287c918b653f292794b4b016d0f6e1ab796aa94d841ad230b6bd396a1"} Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.430763 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" podStartSLOduration=3.4307431409999998 podStartE2EDuration="3.430743141s" podCreationTimestamp="2025-10-07 21:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:12.42898488 +0000 UTC m=+982.467566030" watchObservedRunningTime="2025-10-07 21:12:12.430743141 +0000 UTC m=+982.469324301" Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.514923 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:12 crc kubenswrapper[4926]: E1007 21:12:12.515082 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 21:12:12 crc kubenswrapper[4926]: E1007 21:12:12.515098 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 21:12:12 crc kubenswrapper[4926]: E1007 21:12:12.515141 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift podName:8671e7ac-5c56-4b89-9d72-386b75d12ec1 nodeName:}" failed. No retries permitted until 2025-10-07 21:12:14.515126429 +0000 UTC m=+984.553707579 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift") pod "swift-storage-0" (UID: "8671e7ac-5c56-4b89-9d72-386b75d12ec1") : configmap "swift-ring-files" not found Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.762736 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-n44s7"] Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.763764 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n44s7" Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.772758 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n44s7"] Oct 07 21:12:12 crc kubenswrapper[4926]: I1007 21:12:12.935477 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpjsn\" (UniqueName: \"kubernetes.io/projected/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09-kube-api-access-kpjsn\") pod \"glance-db-create-n44s7\" (UID: \"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09\") " pod="openstack/glance-db-create-n44s7" Oct 07 21:12:13 crc kubenswrapper[4926]: I1007 21:12:13.038633 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpjsn\" (UniqueName: \"kubernetes.io/projected/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09-kube-api-access-kpjsn\") pod \"glance-db-create-n44s7\" (UID: \"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09\") " pod="openstack/glance-db-create-n44s7" Oct 07 21:12:13 crc kubenswrapper[4926]: I1007 21:12:13.071677 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpjsn\" (UniqueName: \"kubernetes.io/projected/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09-kube-api-access-kpjsn\") pod \"glance-db-create-n44s7\" (UID: \"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09\") " pod="openstack/glance-db-create-n44s7" Oct 07 21:12:13 crc kubenswrapper[4926]: I1007 21:12:13.082843 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n44s7" Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.407318 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.443479 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-4rl9j" event={"ID":"51cda538-9dad-4aef-91ac-16af8c38e2dd","Type":"ContainerDied","Data":"1c8f4eb909a86a290a4e91bd2b2834976fec88d50a01d590abbfac4b1f7ddb59"} Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.443518 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c8f4eb909a86a290a4e91bd2b2834976fec88d50a01d590abbfac4b1f7ddb59" Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.443543 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-4rl9j" Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.567120 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tccgh\" (UniqueName: \"kubernetes.io/projected/51cda538-9dad-4aef-91ac-16af8c38e2dd-kube-api-access-tccgh\") pod \"51cda538-9dad-4aef-91ac-16af8c38e2dd\" (UID: \"51cda538-9dad-4aef-91ac-16af8c38e2dd\") " Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.567380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:14 crc kubenswrapper[4926]: E1007 21:12:14.567712 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 21:12:14 crc kubenswrapper[4926]: E1007 21:12:14.567746 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 21:12:14 crc kubenswrapper[4926]: E1007 21:12:14.567806 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift podName:8671e7ac-5c56-4b89-9d72-386b75d12ec1 nodeName:}" failed. No retries permitted until 2025-10-07 21:12:18.567786284 +0000 UTC m=+988.606367434 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift") pod "swift-storage-0" (UID: "8671e7ac-5c56-4b89-9d72-386b75d12ec1") : configmap "swift-ring-files" not found Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.586568 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51cda538-9dad-4aef-91ac-16af8c38e2dd-kube-api-access-tccgh" (OuterVolumeSpecName: "kube-api-access-tccgh") pod "51cda538-9dad-4aef-91ac-16af8c38e2dd" (UID: "51cda538-9dad-4aef-91ac-16af8c38e2dd"). InnerVolumeSpecName "kube-api-access-tccgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:14 crc kubenswrapper[4926]: I1007 21:12:14.669961 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tccgh\" (UniqueName: \"kubernetes.io/projected/51cda538-9dad-4aef-91ac-16af8c38e2dd-kube-api-access-tccgh\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.144822 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-dcmb9"] Oct 07 21:12:17 crc kubenswrapper[4926]: E1007 21:12:17.145965 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51cda538-9dad-4aef-91ac-16af8c38e2dd" containerName="mariadb-database-create" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.145983 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="51cda538-9dad-4aef-91ac-16af8c38e2dd" containerName="mariadb-database-create" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.146253 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="51cda538-9dad-4aef-91ac-16af8c38e2dd" containerName="mariadb-database-create" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.147039 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.153154 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-dcmb9"] Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.328184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6sjg\" (UniqueName: \"kubernetes.io/projected/6bf712ae-977e-409a-901a-13bdac028385-kube-api-access-x6sjg\") pod \"keystone-db-create-dcmb9\" (UID: \"6bf712ae-977e-409a-901a-13bdac028385\") " pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.429541 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6sjg\" (UniqueName: \"kubernetes.io/projected/6bf712ae-977e-409a-901a-13bdac028385-kube-api-access-x6sjg\") pod \"keystone-db-create-dcmb9\" (UID: \"6bf712ae-977e-409a-901a-13bdac028385\") " pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.446574 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6sjg\" (UniqueName: \"kubernetes.io/projected/6bf712ae-977e-409a-901a-13bdac028385-kube-api-access-x6sjg\") pod \"keystone-db-create-dcmb9\" (UID: \"6bf712ae-977e-409a-901a-13bdac028385\") " pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.466793 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-mm68g"] Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.467969 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mm68g" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.469797 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.482601 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mm68g"] Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.634304 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vndlt\" (UniqueName: \"kubernetes.io/projected/4335ef96-e8fe-4f9e-9387-821a589dd37c-kube-api-access-vndlt\") pod \"placement-db-create-mm68g\" (UID: \"4335ef96-e8fe-4f9e-9387-821a589dd37c\") " pod="openstack/placement-db-create-mm68g" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.742186 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vndlt\" (UniqueName: \"kubernetes.io/projected/4335ef96-e8fe-4f9e-9387-821a589dd37c-kube-api-access-vndlt\") pod \"placement-db-create-mm68g\" (UID: \"4335ef96-e8fe-4f9e-9387-821a589dd37c\") " pod="openstack/placement-db-create-mm68g" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.771765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vndlt\" (UniqueName: \"kubernetes.io/projected/4335ef96-e8fe-4f9e-9387-821a589dd37c-kube-api-access-vndlt\") pod \"placement-db-create-mm68g\" (UID: \"4335ef96-e8fe-4f9e-9387-821a589dd37c\") " pod="openstack/placement-db-create-mm68g" Oct 07 21:12:17 crc kubenswrapper[4926]: I1007 21:12:17.818437 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mm68g" Oct 07 21:12:18 crc kubenswrapper[4926]: I1007 21:12:18.021706 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 07 21:12:18 crc kubenswrapper[4926]: I1007 21:12:18.668938 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:18 crc kubenswrapper[4926]: E1007 21:12:18.669265 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 21:12:18 crc kubenswrapper[4926]: E1007 21:12:18.669306 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 21:12:18 crc kubenswrapper[4926]: E1007 21:12:18.669390 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift podName:8671e7ac-5c56-4b89-9d72-386b75d12ec1 nodeName:}" failed. No retries permitted until 2025-10-07 21:12:26.669363205 +0000 UTC m=+996.707944385 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift") pod "swift-storage-0" (UID: "8671e7ac-5c56-4b89-9d72-386b75d12ec1") : configmap "swift-ring-files" not found Oct 07 21:12:19 crc kubenswrapper[4926]: I1007 21:12:19.958544 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:12:20 crc kubenswrapper[4926]: I1007 21:12:20.043986 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9b5c568f-t9jpg"] Oct 07 21:12:20 crc kubenswrapper[4926]: I1007 21:12:20.044384 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="dnsmasq-dns" containerID="cri-o://e4bdac8a8da435e799222ab498f246d4a44a31aa41525ed182770d2dfb8c79e5" gracePeriod=10 Oct 07 21:12:20 crc kubenswrapper[4926]: I1007 21:12:20.516491 4926 generic.go:334] "Generic (PLEG): container finished" podID="869188a9-80d3-454b-8e54-0e885881463f" containerID="e4bdac8a8da435e799222ab498f246d4a44a31aa41525ed182770d2dfb8c79e5" exitCode=0 Oct 07 21:12:20 crc kubenswrapper[4926]: I1007 21:12:20.516563 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" event={"ID":"869188a9-80d3-454b-8e54-0e885881463f","Type":"ContainerDied","Data":"e4bdac8a8da435e799222ab498f246d4a44a31aa41525ed182770d2dfb8c79e5"} Oct 07 21:12:21 crc kubenswrapper[4926]: I1007 21:12:21.530457 4926 generic.go:334] "Generic (PLEG): container finished" podID="5b0d7388-05f4-4df6-b093-c9157a4658b0" containerID="c8c5d091362db91e9aa805fa9aaf324013dc391f8d44fa143c6039bb4fb9af52" exitCode=0 Oct 07 21:12:21 crc kubenswrapper[4926]: I1007 21:12:21.530578 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"5b0d7388-05f4-4df6-b093-c9157a4658b0","Type":"ContainerDied","Data":"c8c5d091362db91e9aa805fa9aaf324013dc391f8d44fa143c6039bb4fb9af52"} Oct 07 21:12:21 crc kubenswrapper[4926]: I1007 21:12:21.535817 4926 generic.go:334] "Generic (PLEG): container finished" podID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerID="aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231" exitCode=0 Oct 07 21:12:21 crc kubenswrapper[4926]: I1007 21:12:21.535876 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57","Type":"ContainerDied","Data":"aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231"} Oct 07 21:12:22 crc kubenswrapper[4926]: I1007 21:12:22.203803 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Oct 07 21:12:22 crc kubenswrapper[4926]: I1007 21:12:22.549415 4926 generic.go:334] "Generic (PLEG): container finished" podID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerID="e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad" exitCode=0 Oct 07 21:12:22 crc kubenswrapper[4926]: I1007 21:12:22.549541 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0","Type":"ContainerDied","Data":"e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad"} Oct 07 21:12:26 crc kubenswrapper[4926]: I1007 21:12:26.739151 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:26 crc kubenswrapper[4926]: E1007 21:12:26.739643 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 21:12:26 crc kubenswrapper[4926]: E1007 21:12:26.739807 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 21:12:26 crc kubenswrapper[4926]: E1007 21:12:26.739897 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift podName:8671e7ac-5c56-4b89-9d72-386b75d12ec1 nodeName:}" failed. No retries permitted until 2025-10-07 21:12:42.739868328 +0000 UTC m=+1012.778449508 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift") pod "swift-storage-0" (UID: "8671e7ac-5c56-4b89-9d72-386b75d12ec1") : configmap "swift-ring-files" not found Oct 07 21:12:27 crc kubenswrapper[4926]: I1007 21:12:27.204102 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Oct 07 21:12:28 crc kubenswrapper[4926]: I1007 21:12:28.856005 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6qkqt" podUID="42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2" containerName="ovn-controller" probeResult="failure" output=< Oct 07 21:12:28 crc kubenswrapper[4926]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 07 21:12:28 crc kubenswrapper[4926]: > Oct 07 21:12:28 crc kubenswrapper[4926]: I1007 21:12:28.883119 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:12:28 crc kubenswrapper[4926]: I1007 21:12:28.886570 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zg8nq" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.207890 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6qkqt-config-5j4pg"] Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.211527 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.218329 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.223569 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6qkqt-config-5j4pg"] Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.400332 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run-ovn\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.400374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-log-ovn\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.400549 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.400585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-scripts\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.400612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwpt\" (UniqueName: \"kubernetes.io/projected/aa84f3d4-b249-4bc3-9796-948f921f2677-kube-api-access-5dwpt\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.400634 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-additional-scripts\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.502273 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dwpt\" (UniqueName: \"kubernetes.io/projected/aa84f3d4-b249-4bc3-9796-948f921f2677-kube-api-access-5dwpt\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.502696 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-additional-scripts\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.502779 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run-ovn\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.502802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-log-ovn\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.502909 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.502949 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-scripts\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.503063 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run-ovn\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.503236 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.503269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-log-ovn\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.503537 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-additional-scripts\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.504818 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-scripts\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.528771 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dwpt\" (UniqueName: \"kubernetes.io/projected/aa84f3d4-b249-4bc3-9796-948f921f2677-kube-api-access-5dwpt\") pod \"ovn-controller-6qkqt-config-5j4pg\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.544091 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.678479 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-f6b6-account-create-hmpr5"] Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.679630 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.682164 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.705104 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-f6b6-account-create-hmpr5"] Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.808612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frfzm\" (UniqueName: \"kubernetes.io/projected/599932ae-c945-4613-9312-484e6a0fce73-kube-api-access-frfzm\") pod \"watcher-f6b6-account-create-hmpr5\" (UID: \"599932ae-c945-4613-9312-484e6a0fce73\") " pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.910620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frfzm\" (UniqueName: \"kubernetes.io/projected/599932ae-c945-4613-9312-484e6a0fce73-kube-api-access-frfzm\") pod \"watcher-f6b6-account-create-hmpr5\" (UID: \"599932ae-c945-4613-9312-484e6a0fce73\") " pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:29 crc kubenswrapper[4926]: I1007 21:12:29.935516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frfzm\" (UniqueName: \"kubernetes.io/projected/599932ae-c945-4613-9312-484e6a0fce73-kube-api-access-frfzm\") pod \"watcher-f6b6-account-create-hmpr5\" (UID: \"599932ae-c945-4613-9312-484e6a0fce73\") " pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.009126 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:30 crc kubenswrapper[4926]: E1007 21:12:30.129028 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00" Oct 07 21:12:30 crc kubenswrapper[4926]: E1007 21:12:30.129311 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus,Image:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00,Command:[],Args:[--web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.enable-remote-write-receiver --web.route-prefix=/ --storage.tsdb.retention.time=24h --storage.tsdb.path=/prometheus --web.config.file=/etc/prometheus/web_config/web-config.yaml],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:web,HostPort:0,ContainerPort:9090,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-out,ReadOnly:true,MountPath:/etc/prometheus/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tls-assets,ReadOnly:true,MountPath:/etc/prometheus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-db,ReadOnly:false,MountPath:/prometheus,SubPath:prometheus-db,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-0,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:web-config,ReadOnly:true,MountPath:/etc/prometheus/web_config/web-config.yaml,SubPath:web-config.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9v28n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/-/healthy,Port:{1 0 web},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:3,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/-/ready,Port:{1 0 web},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:3,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/-/ready,Port:{1 0 web},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:3,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:60,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod prometheus-metric-storage-0_openstack(6ad789bb-88f9-4bec-ba17-46ac114893aa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.635342 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" event={"ID":"869188a9-80d3-454b-8e54-0e885881463f","Type":"ContainerDied","Data":"413c4ed3a617bb5621dc9bb21ef80a2bcc31f7f3970cd0216df1404ac2f0c7fa"} Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.636100 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="413c4ed3a617bb5621dc9bb21ef80a2bcc31f7f3970cd0216df1404ac2f0c7fa" Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.748155 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.884823 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-dcmb9"] Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.896350 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mm68g"] Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.928975 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6tc8\" (UniqueName: \"kubernetes.io/projected/869188a9-80d3-454b-8e54-0e885881463f-kube-api-access-f6tc8\") pod \"869188a9-80d3-454b-8e54-0e885881463f\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.929123 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-nb\") pod \"869188a9-80d3-454b-8e54-0e885881463f\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.929184 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-config\") pod \"869188a9-80d3-454b-8e54-0e885881463f\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.929217 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-sb\") pod \"869188a9-80d3-454b-8e54-0e885881463f\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.929355 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-dns-svc\") pod \"869188a9-80d3-454b-8e54-0e885881463f\" (UID: \"869188a9-80d3-454b-8e54-0e885881463f\") " Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.955145 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869188a9-80d3-454b-8e54-0e885881463f-kube-api-access-f6tc8" (OuterVolumeSpecName: "kube-api-access-f6tc8") pod "869188a9-80d3-454b-8e54-0e885881463f" (UID: "869188a9-80d3-454b-8e54-0e885881463f"). InnerVolumeSpecName "kube-api-access-f6tc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:30 crc kubenswrapper[4926]: I1007 21:12:30.975322 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n44s7"] Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.034247 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6tc8\" (UniqueName: \"kubernetes.io/projected/869188a9-80d3-454b-8e54-0e885881463f-kube-api-access-f6tc8\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.068919 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-config" (OuterVolumeSpecName: "config") pod "869188a9-80d3-454b-8e54-0e885881463f" (UID: "869188a9-80d3-454b-8e54-0e885881463f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.077902 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "869188a9-80d3-454b-8e54-0e885881463f" (UID: "869188a9-80d3-454b-8e54-0e885881463f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.095617 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "869188a9-80d3-454b-8e54-0e885881463f" (UID: "869188a9-80d3-454b-8e54-0e885881463f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.102494 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "869188a9-80d3-454b-8e54-0e885881463f" (UID: "869188a9-80d3-454b-8e54-0e885881463f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.110331 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6qkqt-config-5j4pg"] Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.154242 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.154263 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.154286 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.154294 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869188a9-80d3-454b-8e54-0e885881463f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.195153 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-f6b6-account-create-hmpr5"] Oct 07 21:12:31 crc kubenswrapper[4926]: W1007 21:12:31.206503 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod599932ae_c945_4613_9312_484e6a0fce73.slice/crio-9696c719e0e1f712e8adddec4488c877814d85a1e19c3b5fa5b2f038035ed7a6 WatchSource:0}: Error finding container 9696c719e0e1f712e8adddec4488c877814d85a1e19c3b5fa5b2f038035ed7a6: Status 404 returned error can't find the container with id 9696c719e0e1f712e8adddec4488c877814d85a1e19c3b5fa5b2f038035ed7a6 Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.645895 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57","Type":"ContainerStarted","Data":"d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.647719 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.649155 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mm68g" event={"ID":"4335ef96-e8fe-4f9e-9387-821a589dd37c","Type":"ContainerStarted","Data":"369f76d9fe89d1ab3d8b6ebd6e4dfe1154342268a288df50d47ca932ce7e3202"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.649185 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mm68g" event={"ID":"4335ef96-e8fe-4f9e-9387-821a589dd37c","Type":"ContainerStarted","Data":"0a914947e19e71079bf0b120400e663765eabf7cfc3919fab90f984bfa255f16"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.652616 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6qkqt-config-5j4pg" event={"ID":"aa84f3d4-b249-4bc3-9796-948f921f2677","Type":"ContainerStarted","Data":"5c03254a2d7768cdf760acaea32d64a6b1ffa3d0b3697959a4b8d651472373d8"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.655126 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"5b0d7388-05f4-4df6-b093-c9157a4658b0","Type":"ContainerStarted","Data":"7b486e69260da7a5abfbe07d1f3caa6148dd20eba8f4b8154a0f575c8fab3a3e"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.656034 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.658673 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f6b6-account-create-hmpr5" event={"ID":"599932ae-c945-4613-9312-484e6a0fce73","Type":"ContainerStarted","Data":"628d8bc50b93308b18aeb83208b9bb2b987c35195b274274dfa01631f80bd301"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.658705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f6b6-account-create-hmpr5" event={"ID":"599932ae-c945-4613-9312-484e6a0fce73","Type":"ContainerStarted","Data":"9696c719e0e1f712e8adddec4488c877814d85a1e19c3b5fa5b2f038035ed7a6"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.663701 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dcmb9" event={"ID":"6bf712ae-977e-409a-901a-13bdac028385","Type":"ContainerStarted","Data":"82d52a86993b5f058ce88c7e5bb86b5b19c93a7762d3edca8721930668bac965"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.663751 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dcmb9" event={"ID":"6bf712ae-977e-409a-901a-13bdac028385","Type":"ContainerStarted","Data":"181d2f8c9da9533fc93000a58de51823d30b9447b2c2c970f386b7b07cbbcc94"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.666270 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n44s7" event={"ID":"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09","Type":"ContainerStarted","Data":"81da7eaad690d373488497e682791d3af44fe02b6e4eb4a58cfd54f0a0251897"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.666321 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n44s7" event={"ID":"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09","Type":"ContainerStarted","Data":"939b416894c599a2b5df826b6cf5145c474e4da503fa002d833a2d18fddcba8a"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.677448 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=47.187994549 podStartE2EDuration="1m9.677429806s" podCreationTimestamp="2025-10-07 21:11:22 +0000 UTC" firstStartedPulling="2025-10-07 21:11:24.374464006 +0000 UTC m=+934.413045156" lastFinishedPulling="2025-10-07 21:11:46.863899263 +0000 UTC m=+956.902480413" observedRunningTime="2025-10-07 21:12:31.673831622 +0000 UTC m=+1001.712412792" watchObservedRunningTime="2025-10-07 21:12:31.677429806 +0000 UTC m=+1001.716010966" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.680365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0","Type":"ContainerStarted","Data":"e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.682657 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.685346 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b5c568f-t9jpg" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.685478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fz74j" event={"ID":"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f","Type":"ContainerStarted","Data":"d8f160ad3fc20dd57a97a12ad6a9a21639b3bb3571aa22ea9539071d73a7b528"} Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.711018 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-f6b6-account-create-hmpr5" podStartSLOduration=2.7109975410000002 podStartE2EDuration="2.710997541s" podCreationTimestamp="2025-10-07 21:12:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:31.703983107 +0000 UTC m=+1001.742564257" watchObservedRunningTime="2025-10-07 21:12:31.710997541 +0000 UTC m=+1001.749578691" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.729382 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-n44s7" podStartSLOduration=19.729358393 podStartE2EDuration="19.729358393s" podCreationTimestamp="2025-10-07 21:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:31.720922059 +0000 UTC m=+1001.759503209" watchObservedRunningTime="2025-10-07 21:12:31.729358393 +0000 UTC m=+1001.767939543" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.742443 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-mm68g" podStartSLOduration=14.742428543 podStartE2EDuration="14.742428543s" podCreationTimestamp="2025-10-07 21:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:31.739037444 +0000 UTC m=+1001.777618594" watchObservedRunningTime="2025-10-07 21:12:31.742428543 +0000 UTC m=+1001.781009693" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.757048 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-dcmb9" podStartSLOduration=14.757025366 podStartE2EDuration="14.757025366s" podCreationTimestamp="2025-10-07 21:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:31.755145682 +0000 UTC m=+1001.793726832" watchObservedRunningTime="2025-10-07 21:12:31.757025366 +0000 UTC m=+1001.795606536" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.801568 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=-9223371967.05323 podStartE2EDuration="1m9.801546108s" podCreationTimestamp="2025-10-07 21:11:22 +0000 UTC" firstStartedPulling="2025-10-07 21:11:24.785840234 +0000 UTC m=+934.824421384" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:31.788100388 +0000 UTC m=+1001.826681538" watchObservedRunningTime="2025-10-07 21:12:31.801546108 +0000 UTC m=+1001.840127258" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.805131 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9b5c568f-t9jpg"] Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.812805 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c9b5c568f-t9jpg"] Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.820125 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-fz74j" podStartSLOduration=2.543332685 podStartE2EDuration="20.820109167s" podCreationTimestamp="2025-10-07 21:12:11 +0000 UTC" firstStartedPulling="2025-10-07 21:12:11.974521752 +0000 UTC m=+982.013102902" lastFinishedPulling="2025-10-07 21:12:30.251298234 +0000 UTC m=+1000.289879384" observedRunningTime="2025-10-07 21:12:31.815136442 +0000 UTC m=+1001.853717592" watchObservedRunningTime="2025-10-07 21:12:31.820109167 +0000 UTC m=+1001.858690317" Oct 07 21:12:31 crc kubenswrapper[4926]: I1007 21:12:31.840782 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371967.014011 podStartE2EDuration="1m9.840765246s" podCreationTimestamp="2025-10-07 21:11:22 +0000 UTC" firstStartedPulling="2025-10-07 21:11:24.159982712 +0000 UTC m=+934.198563862" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:31.838735257 +0000 UTC m=+1001.877316407" watchObservedRunningTime="2025-10-07 21:12:31.840765246 +0000 UTC m=+1001.879346396" Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.696166 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869188a9-80d3-454b-8e54-0e885881463f" path="/var/lib/kubelet/pods/869188a9-80d3-454b-8e54-0e885881463f/volumes" Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.697770 4926 generic.go:334] "Generic (PLEG): container finished" podID="6bf712ae-977e-409a-901a-13bdac028385" containerID="82d52a86993b5f058ce88c7e5bb86b5b19c93a7762d3edca8721930668bac965" exitCode=0 Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.697865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dcmb9" event={"ID":"6bf712ae-977e-409a-901a-13bdac028385","Type":"ContainerDied","Data":"82d52a86993b5f058ce88c7e5bb86b5b19c93a7762d3edca8721930668bac965"} Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.700320 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d8c5f01-13f7-4691-b2ec-1539d9e5ee09" containerID="81da7eaad690d373488497e682791d3af44fe02b6e4eb4a58cfd54f0a0251897" exitCode=0 Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.700420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n44s7" event={"ID":"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09","Type":"ContainerDied","Data":"81da7eaad690d373488497e682791d3af44fe02b6e4eb4a58cfd54f0a0251897"} Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.702608 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa84f3d4-b249-4bc3-9796-948f921f2677" containerID="f8ff812f242d17af28a6b92666939f0ec469d943ff421daca201cc8823ac4e5e" exitCode=0 Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.702767 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6qkqt-config-5j4pg" event={"ID":"aa84f3d4-b249-4bc3-9796-948f921f2677","Type":"ContainerDied","Data":"f8ff812f242d17af28a6b92666939f0ec469d943ff421daca201cc8823ac4e5e"} Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.705425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerStarted","Data":"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73"} Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.707447 4926 generic.go:334] "Generic (PLEG): container finished" podID="599932ae-c945-4613-9312-484e6a0fce73" containerID="628d8bc50b93308b18aeb83208b9bb2b987c35195b274274dfa01631f80bd301" exitCode=0 Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.707532 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f6b6-account-create-hmpr5" event={"ID":"599932ae-c945-4613-9312-484e6a0fce73","Type":"ContainerDied","Data":"628d8bc50b93308b18aeb83208b9bb2b987c35195b274274dfa01631f80bd301"} Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.709942 4926 generic.go:334] "Generic (PLEG): container finished" podID="4335ef96-e8fe-4f9e-9387-821a589dd37c" containerID="369f76d9fe89d1ab3d8b6ebd6e4dfe1154342268a288df50d47ca932ce7e3202" exitCode=0 Oct 07 21:12:32 crc kubenswrapper[4926]: I1007 21:12:32.711018 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mm68g" event={"ID":"4335ef96-e8fe-4f9e-9387-821a589dd37c","Type":"ContainerDied","Data":"369f76d9fe89d1ab3d8b6ebd6e4dfe1154342268a288df50d47ca932ce7e3202"} Oct 07 21:12:33 crc kubenswrapper[4926]: I1007 21:12:33.829578 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6qkqt" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.150604 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.204105 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6sjg\" (UniqueName: \"kubernetes.io/projected/6bf712ae-977e-409a-901a-13bdac028385-kube-api-access-x6sjg\") pod \"6bf712ae-977e-409a-901a-13bdac028385\" (UID: \"6bf712ae-977e-409a-901a-13bdac028385\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.213353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf712ae-977e-409a-901a-13bdac028385-kube-api-access-x6sjg" (OuterVolumeSpecName: "kube-api-access-x6sjg") pod "6bf712ae-977e-409a-901a-13bdac028385" (UID: "6bf712ae-977e-409a-901a-13bdac028385"). InnerVolumeSpecName "kube-api-access-x6sjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.305571 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6sjg\" (UniqueName: \"kubernetes.io/projected/6bf712ae-977e-409a-901a-13bdac028385-kube-api-access-x6sjg\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.336664 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mm68g" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.341667 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.408540 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frfzm\" (UniqueName: \"kubernetes.io/projected/599932ae-c945-4613-9312-484e6a0fce73-kube-api-access-frfzm\") pod \"599932ae-c945-4613-9312-484e6a0fce73\" (UID: \"599932ae-c945-4613-9312-484e6a0fce73\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.408603 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vndlt\" (UniqueName: \"kubernetes.io/projected/4335ef96-e8fe-4f9e-9387-821a589dd37c-kube-api-access-vndlt\") pod \"4335ef96-e8fe-4f9e-9387-821a589dd37c\" (UID: \"4335ef96-e8fe-4f9e-9387-821a589dd37c\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.423353 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.442652 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/599932ae-c945-4613-9312-484e6a0fce73-kube-api-access-frfzm" (OuterVolumeSpecName: "kube-api-access-frfzm") pod "599932ae-c945-4613-9312-484e6a0fce73" (UID: "599932ae-c945-4613-9312-484e6a0fce73"). InnerVolumeSpecName "kube-api-access-frfzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.444990 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4335ef96-e8fe-4f9e-9387-821a589dd37c-kube-api-access-vndlt" (OuterVolumeSpecName: "kube-api-access-vndlt") pod "4335ef96-e8fe-4f9e-9387-821a589dd37c" (UID: "4335ef96-e8fe-4f9e-9387-821a589dd37c"). InnerVolumeSpecName "kube-api-access-vndlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.458765 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n44s7" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.510856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run-ovn\") pod \"aa84f3d4-b249-4bc3-9796-948f921f2677\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.510904 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-additional-scripts\") pod \"aa84f3d4-b249-4bc3-9796-948f921f2677\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.510958 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "aa84f3d4-b249-4bc3-9796-948f921f2677" (UID: "aa84f3d4-b249-4bc3-9796-948f921f2677"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.510972 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dwpt\" (UniqueName: \"kubernetes.io/projected/aa84f3d4-b249-4bc3-9796-948f921f2677-kube-api-access-5dwpt\") pod \"aa84f3d4-b249-4bc3-9796-948f921f2677\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511034 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpjsn\" (UniqueName: \"kubernetes.io/projected/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09-kube-api-access-kpjsn\") pod \"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09\" (UID: \"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511082 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-log-ovn\") pod \"aa84f3d4-b249-4bc3-9796-948f921f2677\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511108 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run\") pod \"aa84f3d4-b249-4bc3-9796-948f921f2677\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511123 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-scripts\") pod \"aa84f3d4-b249-4bc3-9796-948f921f2677\" (UID: \"aa84f3d4-b249-4bc3-9796-948f921f2677\") " Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511553 4926 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511572 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frfzm\" (UniqueName: \"kubernetes.io/projected/599932ae-c945-4613-9312-484e6a0fce73-kube-api-access-frfzm\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511582 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vndlt\" (UniqueName: \"kubernetes.io/projected/4335ef96-e8fe-4f9e-9387-821a589dd37c-kube-api-access-vndlt\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.511824 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "aa84f3d4-b249-4bc3-9796-948f921f2677" (UID: "aa84f3d4-b249-4bc3-9796-948f921f2677"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.512282 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run" (OuterVolumeSpecName: "var-run") pod "aa84f3d4-b249-4bc3-9796-948f921f2677" (UID: "aa84f3d4-b249-4bc3-9796-948f921f2677"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.512290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "aa84f3d4-b249-4bc3-9796-948f921f2677" (UID: "aa84f3d4-b249-4bc3-9796-948f921f2677"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.512374 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-scripts" (OuterVolumeSpecName: "scripts") pod "aa84f3d4-b249-4bc3-9796-948f921f2677" (UID: "aa84f3d4-b249-4bc3-9796-948f921f2677"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.514816 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09-kube-api-access-kpjsn" (OuterVolumeSpecName: "kube-api-access-kpjsn") pod "1d8c5f01-13f7-4691-b2ec-1539d9e5ee09" (UID: "1d8c5f01-13f7-4691-b2ec-1539d9e5ee09"). InnerVolumeSpecName "kube-api-access-kpjsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.515542 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa84f3d4-b249-4bc3-9796-948f921f2677-kube-api-access-5dwpt" (OuterVolumeSpecName: "kube-api-access-5dwpt") pod "aa84f3d4-b249-4bc3-9796-948f921f2677" (UID: "aa84f3d4-b249-4bc3-9796-948f921f2677"). InnerVolumeSpecName "kube-api-access-5dwpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.613746 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpjsn\" (UniqueName: \"kubernetes.io/projected/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09-kube-api-access-kpjsn\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.613792 4926 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.613809 4926 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa84f3d4-b249-4bc3-9796-948f921f2677-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.613822 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.613835 4926 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa84f3d4-b249-4bc3-9796-948f921f2677-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.613846 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dwpt\" (UniqueName: \"kubernetes.io/projected/aa84f3d4-b249-4bc3-9796-948f921f2677-kube-api-access-5dwpt\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.727572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n44s7" event={"ID":"1d8c5f01-13f7-4691-b2ec-1539d9e5ee09","Type":"ContainerDied","Data":"939b416894c599a2b5df826b6cf5145c474e4da503fa002d833a2d18fddcba8a"} Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.727674 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="939b416894c599a2b5df826b6cf5145c474e4da503fa002d833a2d18fddcba8a" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.727799 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n44s7" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.733881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6qkqt-config-5j4pg" event={"ID":"aa84f3d4-b249-4bc3-9796-948f921f2677","Type":"ContainerDied","Data":"5c03254a2d7768cdf760acaea32d64a6b1ffa3d0b3697959a4b8d651472373d8"} Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.733918 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c03254a2d7768cdf760acaea32d64a6b1ffa3d0b3697959a4b8d651472373d8" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.733929 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6qkqt-config-5j4pg" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.736639 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mm68g" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.736717 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mm68g" event={"ID":"4335ef96-e8fe-4f9e-9387-821a589dd37c","Type":"ContainerDied","Data":"0a914947e19e71079bf0b120400e663765eabf7cfc3919fab90f984bfa255f16"} Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.736768 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a914947e19e71079bf0b120400e663765eabf7cfc3919fab90f984bfa255f16" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.741021 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f6b6-account-create-hmpr5" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.741075 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f6b6-account-create-hmpr5" event={"ID":"599932ae-c945-4613-9312-484e6a0fce73","Type":"ContainerDied","Data":"9696c719e0e1f712e8adddec4488c877814d85a1e19c3b5fa5b2f038035ed7a6"} Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.741599 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9696c719e0e1f712e8adddec4488c877814d85a1e19c3b5fa5b2f038035ed7a6" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.743945 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dcmb9" event={"ID":"6bf712ae-977e-409a-901a-13bdac028385","Type":"ContainerDied","Data":"181d2f8c9da9533fc93000a58de51823d30b9447b2c2c970f386b7b07cbbcc94"} Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.744009 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="181d2f8c9da9533fc93000a58de51823d30b9447b2c2c970f386b7b07cbbcc94" Oct 07 21:12:34 crc kubenswrapper[4926]: I1007 21:12:34.744168 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dcmb9" Oct 07 21:12:35 crc kubenswrapper[4926]: I1007 21:12:35.527236 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6qkqt-config-5j4pg"] Oct 07 21:12:35 crc kubenswrapper[4926]: I1007 21:12:35.535736 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6qkqt-config-5j4pg"] Oct 07 21:12:36 crc kubenswrapper[4926]: I1007 21:12:36.690938 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa84f3d4-b249-4bc3-9796-948f921f2677" path="/var/lib/kubelet/pods/aa84f3d4-b249-4bc3-9796-948f921f2677/volumes" Oct 07 21:12:37 crc kubenswrapper[4926]: E1007 21:12:37.995611 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/prometheus-metric-storage-0" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" Oct 07 21:12:38 crc kubenswrapper[4926]: I1007 21:12:38.794853 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerStarted","Data":"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c"} Oct 07 21:12:38 crc kubenswrapper[4926]: E1007 21:12:38.797469 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" Oct 07 21:12:39 crc kubenswrapper[4926]: I1007 21:12:39.810152 4926 generic.go:334] "Generic (PLEG): container finished" podID="f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" containerID="d8f160ad3fc20dd57a97a12ad6a9a21639b3bb3571aa22ea9539071d73a7b528" exitCode=0 Oct 07 21:12:39 crc kubenswrapper[4926]: I1007 21:12:39.810243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fz74j" event={"ID":"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f","Type":"ContainerDied","Data":"d8f160ad3fc20dd57a97a12ad6a9a21639b3bb3571aa22ea9539071d73a7b528"} Oct 07 21:12:39 crc kubenswrapper[4926]: E1007 21:12:39.813487 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.179608 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.358675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-dispersionconf\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.358739 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-combined-ca-bundle\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.358822 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-swiftconf\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.358891 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw82m\" (UniqueName: \"kubernetes.io/projected/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-kube-api-access-qw82m\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.358917 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-ring-data-devices\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.358971 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-etc-swift\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.359077 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-scripts\") pod \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\" (UID: \"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f\") " Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.359803 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.359933 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.363883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-kube-api-access-qw82m" (OuterVolumeSpecName: "kube-api-access-qw82m") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "kube-api-access-qw82m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.366497 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.380842 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-scripts" (OuterVolumeSpecName: "scripts") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.382889 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.387326 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" (UID: "f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461400 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461428 4926 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461439 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461450 4926 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461458 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw82m\" (UniqueName: \"kubernetes.io/projected/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-kube-api-access-qw82m\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461466 4926 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.461475 4926 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.840078 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fz74j" event={"ID":"f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f","Type":"ContainerDied","Data":"9b3d19b287c918b653f292794b4b016d0f6e1ab796aa94d841ad230b6bd396a1"} Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.840115 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b3d19b287c918b653f292794b4b016d0f6e1ab796aa94d841ad230b6bd396a1" Oct 07 21:12:41 crc kubenswrapper[4926]: I1007 21:12:41.840246 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fz74j" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.788092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.794660 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8671e7ac-5c56-4b89-9d72-386b75d12ec1-etc-swift\") pod \"swift-storage-0\" (UID: \"8671e7ac-5c56-4b89-9d72-386b75d12ec1\") " pod="openstack/swift-storage-0" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.949412 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.962954 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-2a78-account-create-q726f"] Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963417 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="599932ae-c945-4613-9312-484e6a0fce73" containerName="mariadb-account-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963439 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="599932ae-c945-4613-9312-484e6a0fce73" containerName="mariadb-account-create" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963452 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="init" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963460 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="init" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963471 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf712ae-977e-409a-901a-13bdac028385" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963479 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf712ae-977e-409a-901a-13bdac028385" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963498 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4335ef96-e8fe-4f9e-9387-821a589dd37c" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963506 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4335ef96-e8fe-4f9e-9387-821a589dd37c" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963520 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" containerName="swift-ring-rebalance" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963531 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" containerName="swift-ring-rebalance" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963545 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="dnsmasq-dns" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963555 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="dnsmasq-dns" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963572 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa84f3d4-b249-4bc3-9796-948f921f2677" containerName="ovn-config" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963582 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa84f3d4-b249-4bc3-9796-948f921f2677" containerName="ovn-config" Oct 07 21:12:42 crc kubenswrapper[4926]: E1007 21:12:42.963606 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c5f01-13f7-4691-b2ec-1539d9e5ee09" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963615 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c5f01-13f7-4691-b2ec-1539d9e5ee09" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963801 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="599932ae-c945-4613-9312-484e6a0fce73" containerName="mariadb-account-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963814 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4335ef96-e8fe-4f9e-9387-821a589dd37c" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963829 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8c5f01-13f7-4691-b2ec-1539d9e5ee09" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963849 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f" containerName="swift-ring-rebalance" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963861 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa84f3d4-b249-4bc3-9796-948f921f2677" containerName="ovn-config" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963870 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="869188a9-80d3-454b-8e54-0e885881463f" containerName="dnsmasq-dns" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.963888 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf712ae-977e-409a-901a-13bdac028385" containerName="mariadb-database-create" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.964564 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.967046 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 07 21:12:42 crc kubenswrapper[4926]: I1007 21:12:42.972259 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2a78-account-create-q726f"] Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.095298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d975l\" (UniqueName: \"kubernetes.io/projected/bcbd36c7-e9b4-47e3-82da-03fb48efad6a-kube-api-access-d975l\") pod \"glance-2a78-account-create-q726f\" (UID: \"bcbd36c7-e9b4-47e3-82da-03fb48efad6a\") " pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.197033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d975l\" (UniqueName: \"kubernetes.io/projected/bcbd36c7-e9b4-47e3-82da-03fb48efad6a-kube-api-access-d975l\") pod \"glance-2a78-account-create-q726f\" (UID: \"bcbd36c7-e9b4-47e3-82da-03fb48efad6a\") " pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.223344 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d975l\" (UniqueName: \"kubernetes.io/projected/bcbd36c7-e9b4-47e3-82da-03fb48efad6a-kube-api-access-d975l\") pod \"glance-2a78-account-create-q726f\" (UID: \"bcbd36c7-e9b4-47e3-82da-03fb48efad6a\") " pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.415358 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.574513 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 07 21:12:43 crc kubenswrapper[4926]: W1007 21:12:43.580696 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8671e7ac_5c56_4b89_9d72_386b75d12ec1.slice/crio-4a1b3361be4b1b27cd2927351a1f2622f2026d1eb2c43e1916ee39e972266c32 WatchSource:0}: Error finding container 4a1b3361be4b1b27cd2927351a1f2622f2026d1eb2c43e1916ee39e972266c32: Status 404 returned error can't find the container with id 4a1b3361be4b1b27cd2927351a1f2622f2026d1eb2c43e1916ee39e972266c32 Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.586924 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.864318 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"4a1b3361be4b1b27cd2927351a1f2622f2026d1eb2c43e1916ee39e972266c32"} Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.889525 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 07 21:12:43 crc kubenswrapper[4926]: I1007 21:12:43.908222 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2a78-account-create-q726f"] Oct 07 21:12:43 crc kubenswrapper[4926]: W1007 21:12:43.913025 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcbd36c7_e9b4_47e3_82da_03fb48efad6a.slice/crio-57dd30dfb7fa8157c7db77b386a5fd0f8ec368599bf3486ca1a75c425ddef0ce WatchSource:0}: Error finding container 57dd30dfb7fa8157c7db77b386a5fd0f8ec368599bf3486ca1a75c425ddef0ce: Status 404 returned error can't find the container with id 57dd30dfb7fa8157c7db77b386a5fd0f8ec368599bf3486ca1a75c425ddef0ce Oct 07 21:12:44 crc kubenswrapper[4926]: I1007 21:12:44.272361 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="5b0d7388-05f4-4df6-b093-c9157a4658b0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Oct 07 21:12:44 crc kubenswrapper[4926]: I1007 21:12:44.879508 4926 generic.go:334] "Generic (PLEG): container finished" podID="bcbd36c7-e9b4-47e3-82da-03fb48efad6a" containerID="40deeb91a8cd9dfbd3aa9f20d60a2fb6536d7e4db7279aa9f59ddd4060bf8a82" exitCode=0 Oct 07 21:12:44 crc kubenswrapper[4926]: I1007 21:12:44.879581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2a78-account-create-q726f" event={"ID":"bcbd36c7-e9b4-47e3-82da-03fb48efad6a","Type":"ContainerDied","Data":"40deeb91a8cd9dfbd3aa9f20d60a2fb6536d7e4db7279aa9f59ddd4060bf8a82"} Oct 07 21:12:44 crc kubenswrapper[4926]: I1007 21:12:44.879632 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2a78-account-create-q726f" event={"ID":"bcbd36c7-e9b4-47e3-82da-03fb48efad6a","Type":"ContainerStarted","Data":"57dd30dfb7fa8157c7db77b386a5fd0f8ec368599bf3486ca1a75c425ddef0ce"} Oct 07 21:12:45 crc kubenswrapper[4926]: I1007 21:12:45.894285 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"e7426668d5f81fee56a8146861ca899d000ef79fe409d96928f5463090cda75f"} Oct 07 21:12:45 crc kubenswrapper[4926]: I1007 21:12:45.894642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"9856c0656644d88d9108767a4f0249ceca069948ef51e857eba0d6b77a0bf609"} Oct 07 21:12:45 crc kubenswrapper[4926]: I1007 21:12:45.894653 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"332021c73393fcbdce6ff68ed6f02eef68a8e05696a08b5754a46533660e7d50"} Oct 07 21:12:45 crc kubenswrapper[4926]: I1007 21:12:45.894662 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"690f719af3fa4b61e0bbfec0acda8d8fc36820b407b8618a5625082b2446ae2f"} Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.459929 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.576407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d975l\" (UniqueName: \"kubernetes.io/projected/bcbd36c7-e9b4-47e3-82da-03fb48efad6a-kube-api-access-d975l\") pod \"bcbd36c7-e9b4-47e3-82da-03fb48efad6a\" (UID: \"bcbd36c7-e9b4-47e3-82da-03fb48efad6a\") " Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.594703 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcbd36c7-e9b4-47e3-82da-03fb48efad6a-kube-api-access-d975l" (OuterVolumeSpecName: "kube-api-access-d975l") pod "bcbd36c7-e9b4-47e3-82da-03fb48efad6a" (UID: "bcbd36c7-e9b4-47e3-82da-03fb48efad6a"). InnerVolumeSpecName "kube-api-access-d975l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.678938 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d975l\" (UniqueName: \"kubernetes.io/projected/bcbd36c7-e9b4-47e3-82da-03fb48efad6a-kube-api-access-d975l\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.906617 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2a78-account-create-q726f" event={"ID":"bcbd36c7-e9b4-47e3-82da-03fb48efad6a","Type":"ContainerDied","Data":"57dd30dfb7fa8157c7db77b386a5fd0f8ec368599bf3486ca1a75c425ddef0ce"} Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.907048 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57dd30dfb7fa8157c7db77b386a5fd0f8ec368599bf3486ca1a75c425ddef0ce" Oct 07 21:12:46 crc kubenswrapper[4926]: I1007 21:12:46.906900 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a78-account-create-q726f" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.328908 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2198-account-create-zg64h"] Oct 07 21:12:47 crc kubenswrapper[4926]: E1007 21:12:47.329366 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcbd36c7-e9b4-47e3-82da-03fb48efad6a" containerName="mariadb-account-create" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.329388 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcbd36c7-e9b4-47e3-82da-03fb48efad6a" containerName="mariadb-account-create" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.329771 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcbd36c7-e9b4-47e3-82da-03fb48efad6a" containerName="mariadb-account-create" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.330475 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.333340 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.340813 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2198-account-create-zg64h"] Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.492316 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t5p4\" (UniqueName: \"kubernetes.io/projected/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b-kube-api-access-4t5p4\") pod \"keystone-2198-account-create-zg64h\" (UID: \"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b\") " pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.524325 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9dc0-account-create-xbzb2"] Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.528469 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.536658 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.548440 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9dc0-account-create-xbzb2"] Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.593975 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t5p4\" (UniqueName: \"kubernetes.io/projected/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b-kube-api-access-4t5p4\") pod \"keystone-2198-account-create-zg64h\" (UID: \"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b\") " pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.621332 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t5p4\" (UniqueName: \"kubernetes.io/projected/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b-kube-api-access-4t5p4\") pod \"keystone-2198-account-create-zg64h\" (UID: \"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b\") " pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.659936 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.694935 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rttt6\" (UniqueName: \"kubernetes.io/projected/629a1436-9bd1-4990-8af8-7a54325da442-kube-api-access-rttt6\") pod \"placement-9dc0-account-create-xbzb2\" (UID: \"629a1436-9bd1-4990-8af8-7a54325da442\") " pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.796458 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rttt6\" (UniqueName: \"kubernetes.io/projected/629a1436-9bd1-4990-8af8-7a54325da442-kube-api-access-rttt6\") pod \"placement-9dc0-account-create-xbzb2\" (UID: \"629a1436-9bd1-4990-8af8-7a54325da442\") " pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.818059 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rttt6\" (UniqueName: \"kubernetes.io/projected/629a1436-9bd1-4990-8af8-7a54325da442-kube-api-access-rttt6\") pod \"placement-9dc0-account-create-xbzb2\" (UID: \"629a1436-9bd1-4990-8af8-7a54325da442\") " pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.853626 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.926430 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"53e4fbda2122970be64cac17a01e256ff081cefd67e4e1c12d4138424ead32d4"} Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.927051 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"623f5f3f703182f2db3b237d124a3ced491543622f175a62a2c54d4aa0900638"} Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.927065 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"b7a6f5df183c18ddaeb57b5a7b3966ab18184f6c447345b7e9e26235a06de2cc"} Oct 07 21:12:47 crc kubenswrapper[4926]: I1007 21:12:47.927079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"a748ae4b55f239e9b8522aab8529fbb807a066be6f9079104f8b09d0e1f54b42"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.099921 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-x66gt"] Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.101263 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.105263 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lnm8t" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.105557 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.112734 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-x66gt"] Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.139022 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2198-account-create-zg64h"] Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.202426 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-combined-ca-bundle\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.202489 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-config-data\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.203144 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9fm9\" (UniqueName: \"kubernetes.io/projected/dc48d275-e265-4b1c-985c-bcee19c95284-kube-api-access-k9fm9\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.203433 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-db-sync-config-data\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: W1007 21:12:48.292119 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9d07d5c_a6d6_460d_afac_e30d5c8aff2b.slice/crio-b6184f23bb0301a5ed09e23162a5bf55f30eecd84ef5ddec3cbf97c819e27a30 WatchSource:0}: Error finding container b6184f23bb0301a5ed09e23162a5bf55f30eecd84ef5ddec3cbf97c819e27a30: Status 404 returned error can't find the container with id b6184f23bb0301a5ed09e23162a5bf55f30eecd84ef5ddec3cbf97c819e27a30 Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.305133 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-db-sync-config-data\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.305182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-combined-ca-bundle\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.305233 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-config-data\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.305302 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9fm9\" (UniqueName: \"kubernetes.io/projected/dc48d275-e265-4b1c-985c-bcee19c95284-kube-api-access-k9fm9\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.310165 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-config-data\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.310248 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-combined-ca-bundle\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.310751 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-db-sync-config-data\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.327086 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9fm9\" (UniqueName: \"kubernetes.io/projected/dc48d275-e265-4b1c-985c-bcee19c95284-kube-api-access-k9fm9\") pod \"glance-db-sync-x66gt\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.329111 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9dc0-account-create-xbzb2"] Oct 07 21:12:48 crc kubenswrapper[4926]: W1007 21:12:48.351707 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod629a1436_9bd1_4990_8af8_7a54325da442.slice/crio-2dbee7a4e21cc7ac2b5a14ef8bb917be895beda31d3819e1a1f00659ed7b5f0c WatchSource:0}: Error finding container 2dbee7a4e21cc7ac2b5a14ef8bb917be895beda31d3819e1a1f00659ed7b5f0c: Status 404 returned error can't find the container with id 2dbee7a4e21cc7ac2b5a14ef8bb917be895beda31d3819e1a1f00659ed7b5f0c Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.428445 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x66gt" Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.786045 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-x66gt"] Oct 07 21:12:48 crc kubenswrapper[4926]: W1007 21:12:48.797393 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc48d275_e265_4b1c_985c_bcee19c95284.slice/crio-c43802130b09fe26d818b4d22a562e93944f0b5993bde18fef3d763a2b83dbd4 WatchSource:0}: Error finding container c43802130b09fe26d818b4d22a562e93944f0b5993bde18fef3d763a2b83dbd4: Status 404 returned error can't find the container with id c43802130b09fe26d818b4d22a562e93944f0b5993bde18fef3d763a2b83dbd4 Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.953868 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x66gt" event={"ID":"dc48d275-e265-4b1c-985c-bcee19c95284","Type":"ContainerStarted","Data":"c43802130b09fe26d818b4d22a562e93944f0b5993bde18fef3d763a2b83dbd4"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.959147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"635f65a6750926d7ac87f32a6be4b6e2c13b64ab0fb21e7ed9e274a27e360f15"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.959210 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"bcc84282ea646d06161d1965f1bba93cf6389f2288a0409acd87dc6df090b11a"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.961262 4926 generic.go:334] "Generic (PLEG): container finished" podID="629a1436-9bd1-4990-8af8-7a54325da442" containerID="ac4739a91966a9598e487bdbd6b342e95c5e544d2cd0937a86bc10af0f1c8343" exitCode=0 Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.961322 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9dc0-account-create-xbzb2" event={"ID":"629a1436-9bd1-4990-8af8-7a54325da442","Type":"ContainerDied","Data":"ac4739a91966a9598e487bdbd6b342e95c5e544d2cd0937a86bc10af0f1c8343"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.961341 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9dc0-account-create-xbzb2" event={"ID":"629a1436-9bd1-4990-8af8-7a54325da442","Type":"ContainerStarted","Data":"2dbee7a4e21cc7ac2b5a14ef8bb917be895beda31d3819e1a1f00659ed7b5f0c"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.962559 4926 generic.go:334] "Generic (PLEG): container finished" podID="e9d07d5c-a6d6-460d-afac-e30d5c8aff2b" containerID="d3593fec014128b97a57ec9ef1c247285234c7dadcf1fc711dfffcf2448abc58" exitCode=0 Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.962601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2198-account-create-zg64h" event={"ID":"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b","Type":"ContainerDied","Data":"d3593fec014128b97a57ec9ef1c247285234c7dadcf1fc711dfffcf2448abc58"} Oct 07 21:12:48 crc kubenswrapper[4926]: I1007 21:12:48.962656 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2198-account-create-zg64h" event={"ID":"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b","Type":"ContainerStarted","Data":"b6184f23bb0301a5ed09e23162a5bf55f30eecd84ef5ddec3cbf97c819e27a30"} Oct 07 21:12:49 crc kubenswrapper[4926]: I1007 21:12:49.988208 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"3b2c8cab884467dc25c50ac2dcd7c1804ada26b3bc158431aec74c49bef1bddf"} Oct 07 21:12:49 crc kubenswrapper[4926]: I1007 21:12:49.988692 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"6a7be79c228355e79f7d1a0ca6f657492a4ba1a309059920de98a8e344f6fe5a"} Oct 07 21:12:49 crc kubenswrapper[4926]: I1007 21:12:49.988705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"a3f2a13518b591321b112a2620d056e68ea3f707a79a75bb47a3e1cc1956ee90"} Oct 07 21:12:49 crc kubenswrapper[4926]: I1007 21:12:49.988714 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"56998127759ce28e1ddb780b37a9716a519c0fde8ccfd8f7f14098b7c8abb969"} Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.324938 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.345165 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.442481 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rttt6\" (UniqueName: \"kubernetes.io/projected/629a1436-9bd1-4990-8af8-7a54325da442-kube-api-access-rttt6\") pod \"629a1436-9bd1-4990-8af8-7a54325da442\" (UID: \"629a1436-9bd1-4990-8af8-7a54325da442\") " Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.447356 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629a1436-9bd1-4990-8af8-7a54325da442-kube-api-access-rttt6" (OuterVolumeSpecName: "kube-api-access-rttt6") pod "629a1436-9bd1-4990-8af8-7a54325da442" (UID: "629a1436-9bd1-4990-8af8-7a54325da442"). InnerVolumeSpecName "kube-api-access-rttt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.544421 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t5p4\" (UniqueName: \"kubernetes.io/projected/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b-kube-api-access-4t5p4\") pod \"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b\" (UID: \"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b\") " Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.545160 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rttt6\" (UniqueName: \"kubernetes.io/projected/629a1436-9bd1-4990-8af8-7a54325da442-kube-api-access-rttt6\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.548025 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b-kube-api-access-4t5p4" (OuterVolumeSpecName: "kube-api-access-4t5p4") pod "e9d07d5c-a6d6-460d-afac-e30d5c8aff2b" (UID: "e9d07d5c-a6d6-460d-afac-e30d5c8aff2b"). InnerVolumeSpecName "kube-api-access-4t5p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.646364 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t5p4\" (UniqueName: \"kubernetes.io/projected/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b-kube-api-access-4t5p4\") on node \"crc\" DevicePath \"\"" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.997250 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2198-account-create-zg64h" Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.997251 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2198-account-create-zg64h" event={"ID":"e9d07d5c-a6d6-460d-afac-e30d5c8aff2b","Type":"ContainerDied","Data":"b6184f23bb0301a5ed09e23162a5bf55f30eecd84ef5ddec3cbf97c819e27a30"} Oct 07 21:12:50 crc kubenswrapper[4926]: I1007 21:12:50.997297 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6184f23bb0301a5ed09e23162a5bf55f30eecd84ef5ddec3cbf97c819e27a30" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.004487 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8671e7ac-5c56-4b89-9d72-386b75d12ec1","Type":"ContainerStarted","Data":"b457c058e6e40fe845eab8ea2bd189f97fbca9285a88908dc3a9b5871f540248"} Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.007158 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9dc0-account-create-xbzb2" event={"ID":"629a1436-9bd1-4990-8af8-7a54325da442","Type":"ContainerDied","Data":"2dbee7a4e21cc7ac2b5a14ef8bb917be895beda31d3819e1a1f00659ed7b5f0c"} Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.007182 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dbee7a4e21cc7ac2b5a14ef8bb917be895beda31d3819e1a1f00659ed7b5f0c" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.007249 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9dc0-account-create-xbzb2" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.075389 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.306722926 podStartE2EDuration="42.075365174s" podCreationTimestamp="2025-10-07 21:12:09 +0000 UTC" firstStartedPulling="2025-10-07 21:12:43.583561654 +0000 UTC m=+1013.622142804" lastFinishedPulling="2025-10-07 21:12:48.352203902 +0000 UTC m=+1018.390785052" observedRunningTime="2025-10-07 21:12:51.061480281 +0000 UTC m=+1021.100061531" watchObservedRunningTime="2025-10-07 21:12:51.075365174 +0000 UTC m=+1021.113946324" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.324506 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c6866c67-6cjbf"] Oct 07 21:12:51 crc kubenswrapper[4926]: E1007 21:12:51.324903 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629a1436-9bd1-4990-8af8-7a54325da442" containerName="mariadb-account-create" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.324936 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="629a1436-9bd1-4990-8af8-7a54325da442" containerName="mariadb-account-create" Oct 07 21:12:51 crc kubenswrapper[4926]: E1007 21:12:51.324961 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d07d5c-a6d6-460d-afac-e30d5c8aff2b" containerName="mariadb-account-create" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.324969 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d07d5c-a6d6-460d-afac-e30d5c8aff2b" containerName="mariadb-account-create" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.325158 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d07d5c-a6d6-460d-afac-e30d5c8aff2b" containerName="mariadb-account-create" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.325178 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="629a1436-9bd1-4990-8af8-7a54325da442" containerName="mariadb-account-create" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.326317 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.328989 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.333703 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c6866c67-6cjbf"] Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.461635 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.461727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.462202 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r86vm\" (UniqueName: \"kubernetes.io/projected/69fa95b0-9096-4383-bfbf-3af3c095c9ad-kube-api-access-r86vm\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.462258 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-config\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.462275 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-svc\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.462366 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-swift-storage-0\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.565852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r86vm\" (UniqueName: \"kubernetes.io/projected/69fa95b0-9096-4383-bfbf-3af3c095c9ad-kube-api-access-r86vm\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.565988 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-config\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.566026 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-svc\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.566139 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-swift-storage-0\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.566237 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.566342 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.567132 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-config\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.567475 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-swift-storage-0\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.567538 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-svc\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.568000 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-nb\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.568930 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-sb\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.581734 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r86vm\" (UniqueName: \"kubernetes.io/projected/69fa95b0-9096-4383-bfbf-3af3c095c9ad-kube-api-access-r86vm\") pod \"dnsmasq-dns-5c6866c67-6cjbf\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:51 crc kubenswrapper[4926]: I1007 21:12:51.689656 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:52 crc kubenswrapper[4926]: I1007 21:12:52.137635 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c6866c67-6cjbf"] Oct 07 21:12:52 crc kubenswrapper[4926]: W1007 21:12:52.142580 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69fa95b0_9096_4383_bfbf_3af3c095c9ad.slice/crio-7aa1fb86119ed4a2dae48fefe86e6f9dc95d353056ece663eb238e7981e4bcce WatchSource:0}: Error finding container 7aa1fb86119ed4a2dae48fefe86e6f9dc95d353056ece663eb238e7981e4bcce: Status 404 returned error can't find the container with id 7aa1fb86119ed4a2dae48fefe86e6f9dc95d353056ece663eb238e7981e4bcce Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.028443 4926 generic.go:334] "Generic (PLEG): container finished" podID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerID="4e63e2b1e30f68a064b41b00afdccc785c2b1218fbb90dbcd77b23dbb3a26e49" exitCode=0 Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.028723 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" event={"ID":"69fa95b0-9096-4383-bfbf-3af3c095c9ad","Type":"ContainerDied","Data":"4e63e2b1e30f68a064b41b00afdccc785c2b1218fbb90dbcd77b23dbb3a26e49"} Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.028747 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" event={"ID":"69fa95b0-9096-4383-bfbf-3af3c095c9ad","Type":"ContainerStarted","Data":"7aa1fb86119ed4a2dae48fefe86e6f9dc95d353056ece663eb238e7981e4bcce"} Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.588492 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.890401 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.900143 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rghgq"] Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.901302 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rghgq" Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.913445 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rghgq"] Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.992106 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-g49qh"] Oct 07 21:12:53 crc kubenswrapper[4926]: I1007 21:12:53.993690 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g49qh" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.008825 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpz6j\" (UniqueName: \"kubernetes.io/projected/decadc57-716e-46a0-a0ef-e3c964c8e5b1-kube-api-access-zpz6j\") pod \"barbican-db-create-rghgq\" (UID: \"decadc57-716e-46a0-a0ef-e3c964c8e5b1\") " pod="openstack/barbican-db-create-rghgq" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.013945 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-g49qh"] Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.040106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" event={"ID":"69fa95b0-9096-4383-bfbf-3af3c095c9ad","Type":"ContainerStarted","Data":"3ea737c649028e281a9295bbcc0e840bb716c92299249b00ab2fdfcc847219e7"} Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.040982 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.074808 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" podStartSLOduration=3.074792902 podStartE2EDuration="3.074792902s" podCreationTimestamp="2025-10-07 21:12:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:12:54.065616406 +0000 UTC m=+1024.104197566" watchObservedRunningTime="2025-10-07 21:12:54.074792902 +0000 UTC m=+1024.113374052" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.110237 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95vr6\" (UniqueName: \"kubernetes.io/projected/794c22da-d189-47c3-a776-40f539508653-kube-api-access-95vr6\") pod \"cinder-db-create-g49qh\" (UID: \"794c22da-d189-47c3-a776-40f539508653\") " pod="openstack/cinder-db-create-g49qh" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.110329 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpz6j\" (UniqueName: \"kubernetes.io/projected/decadc57-716e-46a0-a0ef-e3c964c8e5b1-kube-api-access-zpz6j\") pod \"barbican-db-create-rghgq\" (UID: \"decadc57-716e-46a0-a0ef-e3c964c8e5b1\") " pod="openstack/barbican-db-create-rghgq" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.150075 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpz6j\" (UniqueName: \"kubernetes.io/projected/decadc57-716e-46a0-a0ef-e3c964c8e5b1-kube-api-access-zpz6j\") pod \"barbican-db-create-rghgq\" (UID: \"decadc57-716e-46a0-a0ef-e3c964c8e5b1\") " pod="openstack/barbican-db-create-rghgq" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.211950 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95vr6\" (UniqueName: \"kubernetes.io/projected/794c22da-d189-47c3-a776-40f539508653-kube-api-access-95vr6\") pod \"cinder-db-create-g49qh\" (UID: \"794c22da-d189-47c3-a776-40f539508653\") " pod="openstack/cinder-db-create-g49qh" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.218795 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rghgq" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.227478 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-vrdm9"] Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.228546 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.233428 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.233532 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95vr6\" (UniqueName: \"kubernetes.io/projected/794c22da-d189-47c3-a776-40f539508653-kube-api-access-95vr6\") pod \"cinder-db-create-g49qh\" (UID: \"794c22da-d189-47c3-a776-40f539508653\") " pod="openstack/cinder-db-create-g49qh" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.233681 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.233743 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.233941 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-x9gv9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.246163 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vrdm9"] Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.281703 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.298233 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-v5d87"] Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.300705 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v5d87" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.313982 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2ws\" (UniqueName: \"kubernetes.io/projected/1287c912-b6cd-4814-b45e-d557e2aa4f67-kube-api-access-kf2ws\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.314092 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-combined-ca-bundle\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.314164 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-config-data\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.317531 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-v5d87"] Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.341676 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g49qh" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.418696 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwkqt\" (UniqueName: \"kubernetes.io/projected/d7b711cb-48ea-4199-a59e-059d35e747a6-kube-api-access-vwkqt\") pod \"neutron-db-create-v5d87\" (UID: \"d7b711cb-48ea-4199-a59e-059d35e747a6\") " pod="openstack/neutron-db-create-v5d87" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.418743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-config-data\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.418830 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2ws\" (UniqueName: \"kubernetes.io/projected/1287c912-b6cd-4814-b45e-d557e2aa4f67-kube-api-access-kf2ws\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.418896 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-combined-ca-bundle\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.431585 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-combined-ca-bundle\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.434751 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-config-data\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.452234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2ws\" (UniqueName: \"kubernetes.io/projected/1287c912-b6cd-4814-b45e-d557e2aa4f67-kube-api-access-kf2ws\") pod \"keystone-db-sync-vrdm9\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.520814 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwkqt\" (UniqueName: \"kubernetes.io/projected/d7b711cb-48ea-4199-a59e-059d35e747a6-kube-api-access-vwkqt\") pod \"neutron-db-create-v5d87\" (UID: \"d7b711cb-48ea-4199-a59e-059d35e747a6\") " pod="openstack/neutron-db-create-v5d87" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.542996 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwkqt\" (UniqueName: \"kubernetes.io/projected/d7b711cb-48ea-4199-a59e-059d35e747a6-kube-api-access-vwkqt\") pod \"neutron-db-create-v5d87\" (UID: \"d7b711cb-48ea-4199-a59e-059d35e747a6\") " pod="openstack/neutron-db-create-v5d87" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.642995 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.660010 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v5d87" Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.787274 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rghgq"] Oct 07 21:12:54 crc kubenswrapper[4926]: I1007 21:12:54.928909 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-g49qh"] Oct 07 21:12:55 crc kubenswrapper[4926]: I1007 21:12:55.100075 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rghgq" event={"ID":"decadc57-716e-46a0-a0ef-e3c964c8e5b1","Type":"ContainerStarted","Data":"41e4e90eb54f17350227c72256647c1db8edbcc5072f5b7dd0c6dee5afe74c3c"} Oct 07 21:12:55 crc kubenswrapper[4926]: I1007 21:12:55.122237 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g49qh" event={"ID":"794c22da-d189-47c3-a776-40f539508653","Type":"ContainerStarted","Data":"5636ce1cf8dac1bde2102b20b1410e36e87ee0652fff0ec10fc3cbaed2840123"} Oct 07 21:12:55 crc kubenswrapper[4926]: I1007 21:12:55.247170 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-v5d87"] Oct 07 21:12:55 crc kubenswrapper[4926]: I1007 21:12:55.309543 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vrdm9"] Oct 07 21:12:55 crc kubenswrapper[4926]: W1007 21:12:55.657647 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7b711cb_48ea_4199_a59e_059d35e747a6.slice/crio-77b09ff7894307f245ecc9b4fab6bfe633561ffbb41ec5a576a2ecdc70dfd5e8 WatchSource:0}: Error finding container 77b09ff7894307f245ecc9b4fab6bfe633561ffbb41ec5a576a2ecdc70dfd5e8: Status 404 returned error can't find the container with id 77b09ff7894307f245ecc9b4fab6bfe633561ffbb41ec5a576a2ecdc70dfd5e8 Oct 07 21:12:55 crc kubenswrapper[4926]: W1007 21:12:55.660989 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1287c912_b6cd_4814_b45e_d557e2aa4f67.slice/crio-b13c07c74ab1fc747fe98ea760a05b6fc4a5e9bb6e963168fe6356ef0b316e36 WatchSource:0}: Error finding container b13c07c74ab1fc747fe98ea760a05b6fc4a5e9bb6e963168fe6356ef0b316e36: Status 404 returned error can't find the container with id b13c07c74ab1fc747fe98ea760a05b6fc4a5e9bb6e963168fe6356ef0b316e36 Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.137048 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-v5d87" event={"ID":"d7b711cb-48ea-4199-a59e-059d35e747a6","Type":"ContainerStarted","Data":"77b09ff7894307f245ecc9b4fab6bfe633561ffbb41ec5a576a2ecdc70dfd5e8"} Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.138688 4926 generic.go:334] "Generic (PLEG): container finished" podID="decadc57-716e-46a0-a0ef-e3c964c8e5b1" containerID="ad11198ff1d392219f09037153a4f62e80cd8b0c17579626ad76b6e8b94b6d98" exitCode=0 Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.138740 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rghgq" event={"ID":"decadc57-716e-46a0-a0ef-e3c964c8e5b1","Type":"ContainerDied","Data":"ad11198ff1d392219f09037153a4f62e80cd8b0c17579626ad76b6e8b94b6d98"} Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.142085 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vrdm9" event={"ID":"1287c912-b6cd-4814-b45e-d557e2aa4f67","Type":"ContainerStarted","Data":"b13c07c74ab1fc747fe98ea760a05b6fc4a5e9bb6e963168fe6356ef0b316e36"} Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.605998 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-vc4tq"] Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.607398 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.612521 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.612744 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-wpt6l" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.615349 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-vc4tq"] Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.678253 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-db-sync-config-data\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.678323 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-combined-ca-bundle\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.678360 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-config-data\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.678401 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwm8q\" (UniqueName: \"kubernetes.io/projected/c175ad01-09b3-4754-960d-4bc3ef685d2e-kube-api-access-hwm8q\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.780853 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-db-sync-config-data\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.781236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-combined-ca-bundle\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.781293 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-config-data\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.781341 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwm8q\" (UniqueName: \"kubernetes.io/projected/c175ad01-09b3-4754-960d-4bc3ef685d2e-kube-api-access-hwm8q\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.792904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-db-sync-config-data\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.793046 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-combined-ca-bundle\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.796218 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwm8q\" (UniqueName: \"kubernetes.io/projected/c175ad01-09b3-4754-960d-4bc3ef685d2e-kube-api-access-hwm8q\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.798569 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-config-data\") pod \"watcher-db-sync-vc4tq\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:12:56 crc kubenswrapper[4926]: I1007 21:12:56.929865 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:13:01 crc kubenswrapper[4926]: I1007 21:13:01.691304 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:13:01 crc kubenswrapper[4926]: I1007 21:13:01.772387 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67cfd59845-5psqc"] Oct 07 21:13:01 crc kubenswrapper[4926]: I1007 21:13:01.772749 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" podUID="8854b1d1-312d-42cf-b908-182769253998" containerName="dnsmasq-dns" containerID="cri-o://636d612a3231de1ed8a767cb9366b49da87eec095b11e9308ebbe9479aa6e1f0" gracePeriod=10 Oct 07 21:13:02 crc kubenswrapper[4926]: I1007 21:13:02.212260 4926 generic.go:334] "Generic (PLEG): container finished" podID="8854b1d1-312d-42cf-b908-182769253998" containerID="636d612a3231de1ed8a767cb9366b49da87eec095b11e9308ebbe9479aa6e1f0" exitCode=0 Oct 07 21:13:02 crc kubenswrapper[4926]: I1007 21:13:02.212313 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" event={"ID":"8854b1d1-312d-42cf-b908-182769253998","Type":"ContainerDied","Data":"636d612a3231de1ed8a767cb9366b49da87eec095b11e9308ebbe9479aa6e1f0"} Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.150492 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rghgq" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.241356 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rghgq" event={"ID":"decadc57-716e-46a0-a0ef-e3c964c8e5b1","Type":"ContainerDied","Data":"41e4e90eb54f17350227c72256647c1db8edbcc5072f5b7dd0c6dee5afe74c3c"} Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.241516 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41e4e90eb54f17350227c72256647c1db8edbcc5072f5b7dd0c6dee5afe74c3c" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.241575 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rghgq" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.252822 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpz6j\" (UniqueName: \"kubernetes.io/projected/decadc57-716e-46a0-a0ef-e3c964c8e5b1-kube-api-access-zpz6j\") pod \"decadc57-716e-46a0-a0ef-e3c964c8e5b1\" (UID: \"decadc57-716e-46a0-a0ef-e3c964c8e5b1\") " Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.261623 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/decadc57-716e-46a0-a0ef-e3c964c8e5b1-kube-api-access-zpz6j" (OuterVolumeSpecName: "kube-api-access-zpz6j") pod "decadc57-716e-46a0-a0ef-e3c964c8e5b1" (UID: "decadc57-716e-46a0-a0ef-e3c964c8e5b1"). InnerVolumeSpecName "kube-api-access-zpz6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.296039 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.360771 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-nb\") pod \"8854b1d1-312d-42cf-b908-182769253998\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.360856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-config\") pod \"8854b1d1-312d-42cf-b908-182769253998\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.360929 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lpjj\" (UniqueName: \"kubernetes.io/projected/8854b1d1-312d-42cf-b908-182769253998-kube-api-access-8lpjj\") pod \"8854b1d1-312d-42cf-b908-182769253998\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.361031 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-sb\") pod \"8854b1d1-312d-42cf-b908-182769253998\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.361094 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-dns-svc\") pod \"8854b1d1-312d-42cf-b908-182769253998\" (UID: \"8854b1d1-312d-42cf-b908-182769253998\") " Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.361546 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpz6j\" (UniqueName: \"kubernetes.io/projected/decadc57-716e-46a0-a0ef-e3c964c8e5b1-kube-api-access-zpz6j\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.425519 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8854b1d1-312d-42cf-b908-182769253998-kube-api-access-8lpjj" (OuterVolumeSpecName: "kube-api-access-8lpjj") pod "8854b1d1-312d-42cf-b908-182769253998" (UID: "8854b1d1-312d-42cf-b908-182769253998"). InnerVolumeSpecName "kube-api-access-8lpjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.470013 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lpjj\" (UniqueName: \"kubernetes.io/projected/8854b1d1-312d-42cf-b908-182769253998-kube-api-access-8lpjj\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.493048 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8854b1d1-312d-42cf-b908-182769253998" (UID: "8854b1d1-312d-42cf-b908-182769253998"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.500618 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-config" (OuterVolumeSpecName: "config") pod "8854b1d1-312d-42cf-b908-182769253998" (UID: "8854b1d1-312d-42cf-b908-182769253998"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.500914 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8854b1d1-312d-42cf-b908-182769253998" (UID: "8854b1d1-312d-42cf-b908-182769253998"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.523597 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8854b1d1-312d-42cf-b908-182769253998" (UID: "8854b1d1-312d-42cf-b908-182769253998"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.553069 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-vc4tq"] Oct 07 21:13:04 crc kubenswrapper[4926]: W1007 21:13:04.557160 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc175ad01_09b3_4754_960d_4bc3ef685d2e.slice/crio-6f3b409472ad225f7ae57b544b682aea1764bb25ddcda236e819c0072eee6fd9 WatchSource:0}: Error finding container 6f3b409472ad225f7ae57b544b682aea1764bb25ddcda236e819c0072eee6fd9: Status 404 returned error can't find the container with id 6f3b409472ad225f7ae57b544b682aea1764bb25ddcda236e819c0072eee6fd9 Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.571188 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.571227 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.571238 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:04 crc kubenswrapper[4926]: I1007 21:13:04.571247 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8854b1d1-312d-42cf-b908-182769253998-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.264992 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-vc4tq" event={"ID":"c175ad01-09b3-4754-960d-4bc3ef685d2e","Type":"ContainerStarted","Data":"6f3b409472ad225f7ae57b544b682aea1764bb25ddcda236e819c0072eee6fd9"} Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.267071 4926 generic.go:334] "Generic (PLEG): container finished" podID="d7b711cb-48ea-4199-a59e-059d35e747a6" containerID="3113393e5225057a04a2b974c295774883b2e2a8c6d3ce1a4ab71497cc5d967b" exitCode=0 Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.267140 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-v5d87" event={"ID":"d7b711cb-48ea-4199-a59e-059d35e747a6","Type":"ContainerDied","Data":"3113393e5225057a04a2b974c295774883b2e2a8c6d3ce1a4ab71497cc5d967b"} Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.270792 4926 generic.go:334] "Generic (PLEG): container finished" podID="794c22da-d189-47c3-a776-40f539508653" containerID="01d35b191b6388d0a11d93e61a275d4036459aae08e840f566c538d12432e275" exitCode=0 Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.270897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g49qh" event={"ID":"794c22da-d189-47c3-a776-40f539508653","Type":"ContainerDied","Data":"01d35b191b6388d0a11d93e61a275d4036459aae08e840f566c538d12432e275"} Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.274151 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerStarted","Data":"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be"} Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.276399 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x66gt" event={"ID":"dc48d275-e265-4b1c-985c-bcee19c95284","Type":"ContainerStarted","Data":"6f408079d69f154d624584bc67a1257a946b943402c2330c77f3fcfe4add44a8"} Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.283015 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" event={"ID":"8854b1d1-312d-42cf-b908-182769253998","Type":"ContainerDied","Data":"c86f6e125d65e6217d5e71ae91b297645fbc0242999fd07614802c58a6da0c64"} Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.283094 4926 scope.go:117] "RemoveContainer" containerID="636d612a3231de1ed8a767cb9366b49da87eec095b11e9308ebbe9479aa6e1f0" Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.283235 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67cfd59845-5psqc" Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.313123 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.223906244 podStartE2EDuration="1m36.313101939s" podCreationTimestamp="2025-10-07 21:11:29 +0000 UTC" firstStartedPulling="2025-10-07 21:11:48.859211423 +0000 UTC m=+958.897792573" lastFinishedPulling="2025-10-07 21:13:03.948407108 +0000 UTC m=+1033.986988268" observedRunningTime="2025-10-07 21:13:05.307850157 +0000 UTC m=+1035.346431327" watchObservedRunningTime="2025-10-07 21:13:05.313101939 +0000 UTC m=+1035.351683109" Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.325827 4926 scope.go:117] "RemoveContainer" containerID="e4c837174b07a268ad0e0b01ada50cabcfa53b9a6244e4337b0120168f080d64" Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.330396 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-x66gt" podStartSLOduration=1.97354944 podStartE2EDuration="17.33037402s" podCreationTimestamp="2025-10-07 21:12:48 +0000 UTC" firstStartedPulling="2025-10-07 21:12:48.799067149 +0000 UTC m=+1018.837648289" lastFinishedPulling="2025-10-07 21:13:04.155891719 +0000 UTC m=+1034.194472869" observedRunningTime="2025-10-07 21:13:05.322899064 +0000 UTC m=+1035.361480214" watchObservedRunningTime="2025-10-07 21:13:05.33037402 +0000 UTC m=+1035.368955180" Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.388996 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67cfd59845-5psqc"] Oct 07 21:13:05 crc kubenswrapper[4926]: I1007 21:13:05.395887 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67cfd59845-5psqc"] Oct 07 21:13:06 crc kubenswrapper[4926]: I1007 21:13:06.060573 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:06 crc kubenswrapper[4926]: I1007 21:13:06.697030 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8854b1d1-312d-42cf-b908-182769253998" path="/var/lib/kubelet/pods/8854b1d1-312d-42cf-b908-182769253998/volumes" Oct 07 21:13:08 crc kubenswrapper[4926]: I1007 21:13:08.929876 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g49qh" Oct 07 21:13:08 crc kubenswrapper[4926]: I1007 21:13:08.958260 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95vr6\" (UniqueName: \"kubernetes.io/projected/794c22da-d189-47c3-a776-40f539508653-kube-api-access-95vr6\") pod \"794c22da-d189-47c3-a776-40f539508653\" (UID: \"794c22da-d189-47c3-a776-40f539508653\") " Oct 07 21:13:08 crc kubenswrapper[4926]: I1007 21:13:08.966677 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/794c22da-d189-47c3-a776-40f539508653-kube-api-access-95vr6" (OuterVolumeSpecName: "kube-api-access-95vr6") pod "794c22da-d189-47c3-a776-40f539508653" (UID: "794c22da-d189-47c3-a776-40f539508653"). InnerVolumeSpecName "kube-api-access-95vr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:09 crc kubenswrapper[4926]: I1007 21:13:09.067501 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95vr6\" (UniqueName: \"kubernetes.io/projected/794c22da-d189-47c3-a776-40f539508653-kube-api-access-95vr6\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:09 crc kubenswrapper[4926]: I1007 21:13:09.327975 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g49qh" event={"ID":"794c22da-d189-47c3-a776-40f539508653","Type":"ContainerDied","Data":"5636ce1cf8dac1bde2102b20b1410e36e87ee0652fff0ec10fc3cbaed2840123"} Oct 07 21:13:09 crc kubenswrapper[4926]: I1007 21:13:09.328050 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5636ce1cf8dac1bde2102b20b1410e36e87ee0652fff0ec10fc3cbaed2840123" Oct 07 21:13:09 crc kubenswrapper[4926]: I1007 21:13:09.328119 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g49qh" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.982165 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-9ff9-account-create-xkcnw"] Oct 07 21:13:13 crc kubenswrapper[4926]: E1007 21:13:13.983514 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="794c22da-d189-47c3-a776-40f539508653" containerName="mariadb-database-create" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983533 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="794c22da-d189-47c3-a776-40f539508653" containerName="mariadb-database-create" Oct 07 21:13:13 crc kubenswrapper[4926]: E1007 21:13:13.983547 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8854b1d1-312d-42cf-b908-182769253998" containerName="init" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983554 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8854b1d1-312d-42cf-b908-182769253998" containerName="init" Oct 07 21:13:13 crc kubenswrapper[4926]: E1007 21:13:13.983590 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decadc57-716e-46a0-a0ef-e3c964c8e5b1" containerName="mariadb-database-create" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983598 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="decadc57-716e-46a0-a0ef-e3c964c8e5b1" containerName="mariadb-database-create" Oct 07 21:13:13 crc kubenswrapper[4926]: E1007 21:13:13.983615 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8854b1d1-312d-42cf-b908-182769253998" containerName="dnsmasq-dns" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983623 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8854b1d1-312d-42cf-b908-182769253998" containerName="dnsmasq-dns" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983836 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="decadc57-716e-46a0-a0ef-e3c964c8e5b1" containerName="mariadb-database-create" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983861 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="794c22da-d189-47c3-a776-40f539508653" containerName="mariadb-database-create" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.983876 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8854b1d1-312d-42cf-b908-182769253998" containerName="dnsmasq-dns" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.984663 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.987311 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 07 21:13:13 crc kubenswrapper[4926]: I1007 21:13:13.991013 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9ff9-account-create-xkcnw"] Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.064905 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjcj5\" (UniqueName: \"kubernetes.io/projected/534eee94-23c7-4555-809e-9ffdb19b4e66-kube-api-access-tjcj5\") pod \"barbican-9ff9-account-create-xkcnw\" (UID: \"534eee94-23c7-4555-809e-9ffdb19b4e66\") " pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.167369 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjcj5\" (UniqueName: \"kubernetes.io/projected/534eee94-23c7-4555-809e-9ffdb19b4e66-kube-api-access-tjcj5\") pod \"barbican-9ff9-account-create-xkcnw\" (UID: \"534eee94-23c7-4555-809e-9ffdb19b4e66\") " pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.195809 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjcj5\" (UniqueName: \"kubernetes.io/projected/534eee94-23c7-4555-809e-9ffdb19b4e66-kube-api-access-tjcj5\") pod \"barbican-9ff9-account-create-xkcnw\" (UID: \"534eee94-23c7-4555-809e-9ffdb19b4e66\") " pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.316316 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.919115 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v5d87" Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.982320 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwkqt\" (UniqueName: \"kubernetes.io/projected/d7b711cb-48ea-4199-a59e-059d35e747a6-kube-api-access-vwkqt\") pod \"d7b711cb-48ea-4199-a59e-059d35e747a6\" (UID: \"d7b711cb-48ea-4199-a59e-059d35e747a6\") " Oct 07 21:13:14 crc kubenswrapper[4926]: I1007 21:13:14.988334 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b711cb-48ea-4199-a59e-059d35e747a6-kube-api-access-vwkqt" (OuterVolumeSpecName: "kube-api-access-vwkqt") pod "d7b711cb-48ea-4199-a59e-059d35e747a6" (UID: "d7b711cb-48ea-4199-a59e-059d35e747a6"). InnerVolumeSpecName "kube-api-access-vwkqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:15 crc kubenswrapper[4926]: I1007 21:13:15.085171 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwkqt\" (UniqueName: \"kubernetes.io/projected/d7b711cb-48ea-4199-a59e-059d35e747a6-kube-api-access-vwkqt\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:15 crc kubenswrapper[4926]: I1007 21:13:15.387237 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-v5d87" event={"ID":"d7b711cb-48ea-4199-a59e-059d35e747a6","Type":"ContainerDied","Data":"77b09ff7894307f245ecc9b4fab6bfe633561ffbb41ec5a576a2ecdc70dfd5e8"} Oct 07 21:13:15 crc kubenswrapper[4926]: I1007 21:13:15.387692 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77b09ff7894307f245ecc9b4fab6bfe633561ffbb41ec5a576a2ecdc70dfd5e8" Oct 07 21:13:15 crc kubenswrapper[4926]: I1007 21:13:15.387318 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v5d87" Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.061582 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.070254 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.304598 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9ff9-account-create-xkcnw"] Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.398739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vrdm9" event={"ID":"1287c912-b6cd-4814-b45e-d557e2aa4f67","Type":"ContainerStarted","Data":"b74a3ca0df5a079985c18373af3386b833b6bff02f5450fcb731be1dd6c78ef6"} Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.400299 4926 generic.go:334] "Generic (PLEG): container finished" podID="dc48d275-e265-4b1c-985c-bcee19c95284" containerID="6f408079d69f154d624584bc67a1257a946b943402c2330c77f3fcfe4add44a8" exitCode=0 Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.400348 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x66gt" event={"ID":"dc48d275-e265-4b1c-985c-bcee19c95284","Type":"ContainerDied","Data":"6f408079d69f154d624584bc67a1257a946b943402c2330c77f3fcfe4add44a8"} Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.401937 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9ff9-account-create-xkcnw" event={"ID":"534eee94-23c7-4555-809e-9ffdb19b4e66","Type":"ContainerStarted","Data":"dd4d28937ca6442cc02b9540f1a5d28baa5a874f0dccdecb176abbd0b1f1946a"} Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.405798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-vc4tq" event={"ID":"c175ad01-09b3-4754-960d-4bc3ef685d2e","Type":"ContainerStarted","Data":"2a7495928007e508d8643e009b9be3e7f63f20d0375c98e1ad18ae4805d5e38b"} Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.408135 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.426545 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-vrdm9" podStartSLOduration=3.058026031 podStartE2EDuration="22.426513903s" podCreationTimestamp="2025-10-07 21:12:54 +0000 UTC" firstStartedPulling="2025-10-07 21:12:55.669445496 +0000 UTC m=+1025.708026646" lastFinishedPulling="2025-10-07 21:13:15.037933368 +0000 UTC m=+1045.076514518" observedRunningTime="2025-10-07 21:13:16.417574773 +0000 UTC m=+1046.456155923" watchObservedRunningTime="2025-10-07 21:13:16.426513903 +0000 UTC m=+1046.465095083" Oct 07 21:13:16 crc kubenswrapper[4926]: I1007 21:13:16.453634 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-vc4tq" podStartSLOduration=9.488592592 podStartE2EDuration="20.453614549s" podCreationTimestamp="2025-10-07 21:12:56 +0000 UTC" firstStartedPulling="2025-10-07 21:13:04.559529562 +0000 UTC m=+1034.598110712" lastFinishedPulling="2025-10-07 21:13:15.524551509 +0000 UTC m=+1045.563132669" observedRunningTime="2025-10-07 21:13:16.450935061 +0000 UTC m=+1046.489516201" watchObservedRunningTime="2025-10-07 21:13:16.453614549 +0000 UTC m=+1046.492195699" Oct 07 21:13:17 crc kubenswrapper[4926]: I1007 21:13:17.414560 4926 generic.go:334] "Generic (PLEG): container finished" podID="534eee94-23c7-4555-809e-9ffdb19b4e66" containerID="7a82c2fc0970713984e825babfab5a475d63c086259872c9c94b7efb4957cda0" exitCode=0 Oct 07 21:13:17 crc kubenswrapper[4926]: I1007 21:13:17.414726 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9ff9-account-create-xkcnw" event={"ID":"534eee94-23c7-4555-809e-9ffdb19b4e66","Type":"ContainerDied","Data":"7a82c2fc0970713984e825babfab5a475d63c086259872c9c94b7efb4957cda0"} Oct 07 21:13:17 crc kubenswrapper[4926]: I1007 21:13:17.883315 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x66gt" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.064934 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-config-data\") pod \"dc48d275-e265-4b1c-985c-bcee19c95284\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.065008 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-db-sync-config-data\") pod \"dc48d275-e265-4b1c-985c-bcee19c95284\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.065048 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-combined-ca-bundle\") pod \"dc48d275-e265-4b1c-985c-bcee19c95284\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.065096 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9fm9\" (UniqueName: \"kubernetes.io/projected/dc48d275-e265-4b1c-985c-bcee19c95284-kube-api-access-k9fm9\") pod \"dc48d275-e265-4b1c-985c-bcee19c95284\" (UID: \"dc48d275-e265-4b1c-985c-bcee19c95284\") " Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.072321 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dc48d275-e265-4b1c-985c-bcee19c95284" (UID: "dc48d275-e265-4b1c-985c-bcee19c95284"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.077392 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc48d275-e265-4b1c-985c-bcee19c95284-kube-api-access-k9fm9" (OuterVolumeSpecName: "kube-api-access-k9fm9") pod "dc48d275-e265-4b1c-985c-bcee19c95284" (UID: "dc48d275-e265-4b1c-985c-bcee19c95284"). InnerVolumeSpecName "kube-api-access-k9fm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.103630 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc48d275-e265-4b1c-985c-bcee19c95284" (UID: "dc48d275-e265-4b1c-985c-bcee19c95284"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.136512 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-config-data" (OuterVolumeSpecName: "config-data") pod "dc48d275-e265-4b1c-985c-bcee19c95284" (UID: "dc48d275-e265-4b1c-985c-bcee19c95284"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.166985 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.167027 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.167040 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc48d275-e265-4b1c-985c-bcee19c95284-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.167053 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9fm9\" (UniqueName: \"kubernetes.io/projected/dc48d275-e265-4b1c-985c-bcee19c95284-kube-api-access-k9fm9\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.438459 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x66gt" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.438457 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x66gt" event={"ID":"dc48d275-e265-4b1c-985c-bcee19c95284","Type":"ContainerDied","Data":"c43802130b09fe26d818b4d22a562e93944f0b5993bde18fef3d763a2b83dbd4"} Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.438611 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c43802130b09fe26d818b4d22a562e93944f0b5993bde18fef3d763a2b83dbd4" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.797453 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.828772 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5db7bcf759-7zm8s"] Oct 07 21:13:18 crc kubenswrapper[4926]: E1007 21:13:18.829182 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b711cb-48ea-4199-a59e-059d35e747a6" containerName="mariadb-database-create" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.829207 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b711cb-48ea-4199-a59e-059d35e747a6" containerName="mariadb-database-create" Oct 07 21:13:18 crc kubenswrapper[4926]: E1007 21:13:18.829231 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc48d275-e265-4b1c-985c-bcee19c95284" containerName="glance-db-sync" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.829237 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc48d275-e265-4b1c-985c-bcee19c95284" containerName="glance-db-sync" Oct 07 21:13:18 crc kubenswrapper[4926]: E1007 21:13:18.829266 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534eee94-23c7-4555-809e-9ffdb19b4e66" containerName="mariadb-account-create" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.829273 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="534eee94-23c7-4555-809e-9ffdb19b4e66" containerName="mariadb-account-create" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.829441 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="534eee94-23c7-4555-809e-9ffdb19b4e66" containerName="mariadb-account-create" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.829455 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b711cb-48ea-4199-a59e-059d35e747a6" containerName="mariadb-database-create" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.829465 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc48d275-e265-4b1c-985c-bcee19c95284" containerName="glance-db-sync" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.830435 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.854319 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5db7bcf759-7zm8s"] Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982234 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjcj5\" (UniqueName: \"kubernetes.io/projected/534eee94-23c7-4555-809e-9ffdb19b4e66-kube-api-access-tjcj5\") pod \"534eee94-23c7-4555-809e-9ffdb19b4e66\" (UID: \"534eee94-23c7-4555-809e-9ffdb19b4e66\") " Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-config\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982652 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-nb\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982670 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-svc\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982693 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-sb\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982716 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-swift-storage-0\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.982753 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpqgq\" (UniqueName: \"kubernetes.io/projected/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-kube-api-access-cpqgq\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:18 crc kubenswrapper[4926]: I1007 21:13:18.999360 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534eee94-23c7-4555-809e-9ffdb19b4e66-kube-api-access-tjcj5" (OuterVolumeSpecName: "kube-api-access-tjcj5") pod "534eee94-23c7-4555-809e-9ffdb19b4e66" (UID: "534eee94-23c7-4555-809e-9ffdb19b4e66"). InnerVolumeSpecName "kube-api-access-tjcj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpqgq\" (UniqueName: \"kubernetes.io/projected/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-kube-api-access-cpqgq\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086544 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-config\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086601 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-nb\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086632 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-svc\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086665 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-sb\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086699 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-swift-storage-0\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.086765 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjcj5\" (UniqueName: \"kubernetes.io/projected/534eee94-23c7-4555-809e-9ffdb19b4e66-kube-api-access-tjcj5\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.087841 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-swift-storage-0\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.087885 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-svc\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.087911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-sb\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.087875 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-config\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.087965 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-nb\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.117886 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpqgq\" (UniqueName: \"kubernetes.io/projected/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-kube-api-access-cpqgq\") pod \"dnsmasq-dns-5db7bcf759-7zm8s\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.178482 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.336650 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.337183 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="config-reloader" containerID="cri-o://06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" gracePeriod=600 Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.337434 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="thanos-sidecar" containerID="cri-o://da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" gracePeriod=600 Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.337449 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="prometheus" containerID="cri-o://6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" gracePeriod=600 Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.464833 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9ff9-account-create-xkcnw" event={"ID":"534eee94-23c7-4555-809e-9ffdb19b4e66","Type":"ContainerDied","Data":"dd4d28937ca6442cc02b9540f1a5d28baa5a874f0dccdecb176abbd0b1f1946a"} Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.464888 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd4d28937ca6442cc02b9540f1a5d28baa5a874f0dccdecb176abbd0b1f1946a" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.464968 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9ff9-account-create-xkcnw" Oct 07 21:13:19 crc kubenswrapper[4926]: I1007 21:13:19.659038 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5db7bcf759-7zm8s"] Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.076649 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.212936 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213005 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ad789bb-88f9-4bec-ba17-46ac114893aa-config-out\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213041 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-config\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213092 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ad789bb-88f9-4bec-ba17-46ac114893aa-prometheus-metric-storage-rulefiles-0\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213122 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-web-config\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213150 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v28n\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-kube-api-access-9v28n\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213227 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-thanos-prometheus-http-client-file\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.213311 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-tls-assets\") pod \"6ad789bb-88f9-4bec-ba17-46ac114893aa\" (UID: \"6ad789bb-88f9-4bec-ba17-46ac114893aa\") " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.214240 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ad789bb-88f9-4bec-ba17-46ac114893aa-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.219919 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.219966 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-kube-api-access-9v28n" (OuterVolumeSpecName: "kube-api-access-9v28n") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "kube-api-access-9v28n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.220420 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-config" (OuterVolumeSpecName: "config") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.226946 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.226972 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ad789bb-88f9-4bec-ba17-46ac114893aa-config-out" (OuterVolumeSpecName: "config-out") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.265250 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-web-config" (OuterVolumeSpecName: "web-config") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.265460 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "6ad789bb-88f9-4bec-ba17-46ac114893aa" (UID: "6ad789bb-88f9-4bec-ba17-46ac114893aa"). InnerVolumeSpecName "pvc-25bd2660-6460-4b17-9118-3d7559e7645c". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316079 4926 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316127 4926 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316167 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") on node \"crc\" " Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316182 4926 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ad789bb-88f9-4bec-ba17-46ac114893aa-config-out\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316209 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316256 4926 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ad789bb-88f9-4bec-ba17-46ac114893aa-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316269 4926 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ad789bb-88f9-4bec-ba17-46ac114893aa-web-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.316281 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v28n\" (UniqueName: \"kubernetes.io/projected/6ad789bb-88f9-4bec-ba17-46ac114893aa-kube-api-access-9v28n\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.344615 4926 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.344792 4926 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-25bd2660-6460-4b17-9118-3d7559e7645c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c") on node "crc" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.416913 4926 reconciler_common.go:293] "Volume detached for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477296 4926 generic.go:334] "Generic (PLEG): container finished" podID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerID="6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" exitCode=0 Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477324 4926 generic.go:334] "Generic (PLEG): container finished" podID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerID="da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" exitCode=0 Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477333 4926 generic.go:334] "Generic (PLEG): container finished" podID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerID="06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" exitCode=0 Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477378 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerDied","Data":"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477406 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerDied","Data":"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477416 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerDied","Data":"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477427 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ad789bb-88f9-4bec-ba17-46ac114893aa","Type":"ContainerDied","Data":"5ffeef0a014c372039674a192f6a48d3a27f9f44a2b284f43f137b784b4c49df"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477448 4926 scope.go:117] "RemoveContainer" containerID="6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.477461 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.483762 4926 generic.go:334] "Generic (PLEG): container finished" podID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerID="9fdff9bfbc66f22cd924bbc3adc3ee8f4b86de87aad4ffe11be6ab776a44d092" exitCode=0 Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.483972 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" event={"ID":"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6","Type":"ContainerDied","Data":"9fdff9bfbc66f22cd924bbc3adc3ee8f4b86de87aad4ffe11be6ab776a44d092"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.484032 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" event={"ID":"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6","Type":"ContainerStarted","Data":"6da72074707ec2106eeccd49fe846e3c0da7176cafffeafbf08b76c1ef77a84c"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.498167 4926 generic.go:334] "Generic (PLEG): container finished" podID="c175ad01-09b3-4754-960d-4bc3ef685d2e" containerID="2a7495928007e508d8643e009b9be3e7f63f20d0375c98e1ad18ae4805d5e38b" exitCode=0 Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.498257 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-vc4tq" event={"ID":"c175ad01-09b3-4754-960d-4bc3ef685d2e","Type":"ContainerDied","Data":"2a7495928007e508d8643e009b9be3e7f63f20d0375c98e1ad18ae4805d5e38b"} Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.506711 4926 scope.go:117] "RemoveContainer" containerID="da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.565951 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.566845 4926 scope.go:117] "RemoveContainer" containerID="06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.578345 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.586402 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.586899 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="init-config-reloader" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.586917 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="init-config-reloader" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.586931 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="prometheus" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.586938 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="prometheus" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.586952 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="thanos-sidecar" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.586960 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="thanos-sidecar" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.586989 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="config-reloader" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.586995 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="config-reloader" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.587165 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="config-reloader" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.587184 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="prometheus" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.587229 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" containerName="thanos-sidecar" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.588864 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.591084 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.592274 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hckhp" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.592514 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.592953 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.597001 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.597135 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.600086 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.601707 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.602702 4926 scope.go:117] "RemoveContainer" containerID="572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.651154 4926 scope.go:117] "RemoveContainer" containerID="6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.651601 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": container with ID starting with 6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be not found: ID does not exist" containerID="6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.651636 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be"} err="failed to get container status \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": rpc error: code = NotFound desc = could not find container \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": container with ID starting with 6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.651685 4926 scope.go:117] "RemoveContainer" containerID="da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.652661 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": container with ID starting with da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c not found: ID does not exist" containerID="da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.652695 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c"} err="failed to get container status \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": rpc error: code = NotFound desc = could not find container \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": container with ID starting with da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.652719 4926 scope.go:117] "RemoveContainer" containerID="06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.653005 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": container with ID starting with 06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73 not found: ID does not exist" containerID="06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653028 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73"} err="failed to get container status \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": rpc error: code = NotFound desc = could not find container \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": container with ID starting with 06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73 not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653042 4926 scope.go:117] "RemoveContainer" containerID="572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08" Oct 07 21:13:20 crc kubenswrapper[4926]: E1007 21:13:20.653281 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": container with ID starting with 572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08 not found: ID does not exist" containerID="572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653300 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08"} err="failed to get container status \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": rpc error: code = NotFound desc = could not find container \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": container with ID starting with 572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08 not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653312 4926 scope.go:117] "RemoveContainer" containerID="6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653461 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be"} err="failed to get container status \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": rpc error: code = NotFound desc = could not find container \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": container with ID starting with 6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653476 4926 scope.go:117] "RemoveContainer" containerID="da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653618 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c"} err="failed to get container status \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": rpc error: code = NotFound desc = could not find container \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": container with ID starting with da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653632 4926 scope.go:117] "RemoveContainer" containerID="06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653814 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73"} err="failed to get container status \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": rpc error: code = NotFound desc = could not find container \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": container with ID starting with 06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73 not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.653836 4926 scope.go:117] "RemoveContainer" containerID="572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654034 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08"} err="failed to get container status \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": rpc error: code = NotFound desc = could not find container \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": container with ID starting with 572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08 not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654052 4926 scope.go:117] "RemoveContainer" containerID="6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654215 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be"} err="failed to get container status \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": rpc error: code = NotFound desc = could not find container \"6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be\": container with ID starting with 6ea9703bc74fafb725692f7c515253a38791d5b24376bf98e2955bbd9d2728be not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654234 4926 scope.go:117] "RemoveContainer" containerID="da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654430 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c"} err="failed to get container status \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": rpc error: code = NotFound desc = could not find container \"da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c\": container with ID starting with da208f09f1ba63d8795a79058f6a77911983a1dd51a8bcaf9fc6b059322d1d6c not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654453 4926 scope.go:117] "RemoveContainer" containerID="06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654647 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73"} err="failed to get container status \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": rpc error: code = NotFound desc = could not find container \"06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73\": container with ID starting with 06495ff187975e9ca9e6d2a4910355a133d8c7d9e9294106b998ef16af159b73 not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.654669 4926 scope.go:117] "RemoveContainer" containerID="572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.655248 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08"} err="failed to get container status \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": rpc error: code = NotFound desc = could not find container \"572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08\": container with ID starting with 572039c91f004bf454cc9472202631ea8a6319bc47731cae780202de173efa08 not found: ID does not exist" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.690452 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ad789bb-88f9-4bec-ba17-46ac114893aa" path="/var/lib/kubelet/pods/6ad789bb-88f9-4bec-ba17-46ac114893aa/volumes" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720672 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7db60520-2468-44b2-9ca3-99c1bfe04bad-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720720 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7db60520-2468-44b2-9ca3-99c1bfe04bad-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720778 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nllkk\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-kube-api-access-nllkk\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720810 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720832 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720849 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720883 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720902 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720921 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720942 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.720994 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-config\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822666 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822716 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822761 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822815 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822880 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.822989 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-config\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.823047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7db60520-2468-44b2-9ca3-99c1bfe04bad-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.823084 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7db60520-2468-44b2-9ca3-99c1bfe04bad-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.823172 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nllkk\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-kube-api-access-nllkk\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.824631 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7db60520-2468-44b2-9ca3-99c1bfe04bad-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.827003 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-config\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.827226 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.827421 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7db60520-2468-44b2-9ca3-99c1bfe04bad-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.828185 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.830032 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.835851 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.836766 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.836973 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.837930 4926 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.837963 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/981924a86d194783a644e1657994993df571041fb46e6d3f243affb52a3a59ba/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.849406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nllkk\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-kube-api-access-nllkk\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.876861 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:20 crc kubenswrapper[4926]: I1007 21:13:20.940519 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:21 crc kubenswrapper[4926]: I1007 21:13:21.508535 4926 generic.go:334] "Generic (PLEG): container finished" podID="1287c912-b6cd-4814-b45e-d557e2aa4f67" containerID="b74a3ca0df5a079985c18373af3386b833b6bff02f5450fcb731be1dd6c78ef6" exitCode=0 Oct 07 21:13:21 crc kubenswrapper[4926]: I1007 21:13:21.508642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vrdm9" event={"ID":"1287c912-b6cd-4814-b45e-d557e2aa4f67","Type":"ContainerDied","Data":"b74a3ca0df5a079985c18373af3386b833b6bff02f5450fcb731be1dd6c78ef6"} Oct 07 21:13:21 crc kubenswrapper[4926]: I1007 21:13:21.513345 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" event={"ID":"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6","Type":"ContainerStarted","Data":"82dd6763f4f094f3425331e1a229a95a8393fc34426d5d0ee8dc099bf7747414"} Oct 07 21:13:21 crc kubenswrapper[4926]: I1007 21:13:21.513502 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:21 crc kubenswrapper[4926]: I1007 21:13:21.593274 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" podStartSLOduration=3.593245142 podStartE2EDuration="3.593245142s" podCreationTimestamp="2025-10-07 21:13:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:21.561347397 +0000 UTC m=+1051.599928547" watchObservedRunningTime="2025-10-07 21:13:21.593245142 +0000 UTC m=+1051.631826322" Oct 07 21:13:21 crc kubenswrapper[4926]: I1007 21:13:21.613232 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.063002 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.169176 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-combined-ca-bundle\") pod \"c175ad01-09b3-4754-960d-4bc3ef685d2e\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.169269 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwm8q\" (UniqueName: \"kubernetes.io/projected/c175ad01-09b3-4754-960d-4bc3ef685d2e-kube-api-access-hwm8q\") pod \"c175ad01-09b3-4754-960d-4bc3ef685d2e\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.169365 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-db-sync-config-data\") pod \"c175ad01-09b3-4754-960d-4bc3ef685d2e\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.170046 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-config-data\") pod \"c175ad01-09b3-4754-960d-4bc3ef685d2e\" (UID: \"c175ad01-09b3-4754-960d-4bc3ef685d2e\") " Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.178185 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c175ad01-09b3-4754-960d-4bc3ef685d2e" (UID: "c175ad01-09b3-4754-960d-4bc3ef685d2e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.179539 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c175ad01-09b3-4754-960d-4bc3ef685d2e-kube-api-access-hwm8q" (OuterVolumeSpecName: "kube-api-access-hwm8q") pod "c175ad01-09b3-4754-960d-4bc3ef685d2e" (UID: "c175ad01-09b3-4754-960d-4bc3ef685d2e"). InnerVolumeSpecName "kube-api-access-hwm8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.204734 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c175ad01-09b3-4754-960d-4bc3ef685d2e" (UID: "c175ad01-09b3-4754-960d-4bc3ef685d2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.224905 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-config-data" (OuterVolumeSpecName: "config-data") pod "c175ad01-09b3-4754-960d-4bc3ef685d2e" (UID: "c175ad01-09b3-4754-960d-4bc3ef685d2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.272137 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.272180 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.272215 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwm8q\" (UniqueName: \"kubernetes.io/projected/c175ad01-09b3-4754-960d-4bc3ef685d2e-kube-api-access-hwm8q\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.272229 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c175ad01-09b3-4754-960d-4bc3ef685d2e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.524081 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-vc4tq" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.524086 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-vc4tq" event={"ID":"c175ad01-09b3-4754-960d-4bc3ef685d2e","Type":"ContainerDied","Data":"6f3b409472ad225f7ae57b544b682aea1764bb25ddcda236e819c0072eee6fd9"} Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.524146 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f3b409472ad225f7ae57b544b682aea1764bb25ddcda236e819c0072eee6fd9" Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.526557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerStarted","Data":"4a1eb84374483306bb63ca335c378799246b25a506fd0f62948d8a5c0574aa57"} Oct 07 21:13:22 crc kubenswrapper[4926]: I1007 21:13:22.981899 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.093538 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf2ws\" (UniqueName: \"kubernetes.io/projected/1287c912-b6cd-4814-b45e-d557e2aa4f67-kube-api-access-kf2ws\") pod \"1287c912-b6cd-4814-b45e-d557e2aa4f67\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.093987 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-config-data\") pod \"1287c912-b6cd-4814-b45e-d557e2aa4f67\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.094109 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-combined-ca-bundle\") pod \"1287c912-b6cd-4814-b45e-d557e2aa4f67\" (UID: \"1287c912-b6cd-4814-b45e-d557e2aa4f67\") " Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.120941 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1287c912-b6cd-4814-b45e-d557e2aa4f67-kube-api-access-kf2ws" (OuterVolumeSpecName: "kube-api-access-kf2ws") pod "1287c912-b6cd-4814-b45e-d557e2aa4f67" (UID: "1287c912-b6cd-4814-b45e-d557e2aa4f67"). InnerVolumeSpecName "kube-api-access-kf2ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.161606 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-config-data" (OuterVolumeSpecName: "config-data") pod "1287c912-b6cd-4814-b45e-d557e2aa4f67" (UID: "1287c912-b6cd-4814-b45e-d557e2aa4f67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.164909 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1287c912-b6cd-4814-b45e-d557e2aa4f67" (UID: "1287c912-b6cd-4814-b45e-d557e2aa4f67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.196174 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.196260 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1287c912-b6cd-4814-b45e-d557e2aa4f67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.196273 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf2ws\" (UniqueName: \"kubernetes.io/projected/1287c912-b6cd-4814-b45e-d557e2aa4f67-kube-api-access-kf2ws\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.536411 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vrdm9" event={"ID":"1287c912-b6cd-4814-b45e-d557e2aa4f67","Type":"ContainerDied","Data":"b13c07c74ab1fc747fe98ea760a05b6fc4a5e9bb6e963168fe6356ef0b316e36"} Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.536452 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b13c07c74ab1fc747fe98ea760a05b6fc4a5e9bb6e963168fe6356ef0b316e36" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.536514 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vrdm9" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.762355 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db7bcf759-7zm8s"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.763008 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerName="dnsmasq-dns" containerID="cri-o://82dd6763f4f094f3425331e1a229a95a8393fc34426d5d0ee8dc099bf7747414" gracePeriod=10 Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.801873 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-z5v4c"] Oct 07 21:13:23 crc kubenswrapper[4926]: E1007 21:13:23.802322 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c175ad01-09b3-4754-960d-4bc3ef685d2e" containerName="watcher-db-sync" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.802339 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c175ad01-09b3-4754-960d-4bc3ef685d2e" containerName="watcher-db-sync" Oct 07 21:13:23 crc kubenswrapper[4926]: E1007 21:13:23.802362 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1287c912-b6cd-4814-b45e-d557e2aa4f67" containerName="keystone-db-sync" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.802368 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1287c912-b6cd-4814-b45e-d557e2aa4f67" containerName="keystone-db-sync" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.802532 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c175ad01-09b3-4754-960d-4bc3ef685d2e" containerName="watcher-db-sync" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.802550 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1287c912-b6cd-4814-b45e-d557e2aa4f67" containerName="keystone-db-sync" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.803143 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.809818 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-x9gv9" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.810057 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.810184 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.810315 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.821137 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f5f8c579-4ls4k"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.829953 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.849968 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-z5v4c"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.905701 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5f8c579-4ls4k"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.917337 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-credential-keys\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.917552 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-scripts\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.917619 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-fernet-keys\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.917702 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nhxw\" (UniqueName: \"kubernetes.io/projected/920e2024-0025-4d13-a17c-a6993cb44eda-kube-api-access-4nhxw\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.917822 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-combined-ca-bundle\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.917969 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-config-data\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.918040 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.919340 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.926836 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-wpt6l" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.943065 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.944296 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.953487 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.953558 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.970476 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.976742 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.997018 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:23 crc kubenswrapper[4926]: I1007 21:13:23.999920 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.001960 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.048504 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-credential-keys\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.048617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-fernet-keys\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.048684 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-scripts\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.048794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nhxw\" (UniqueName: \"kubernetes.io/projected/920e2024-0025-4d13-a17c-a6993cb44eda-kube-api-access-4nhxw\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.048900 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-config\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049169 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-combined-ca-bundle\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049344 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049440 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-config-data\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049503 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfv55\" (UniqueName: \"kubernetes.io/projected/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-kube-api-access-mfv55\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.049661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-svc\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.055483 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-scripts\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.084155 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-combined-ca-bundle\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.089046 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-config-data\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.139883 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.150791 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nhxw\" (UniqueName: \"kubernetes.io/projected/920e2024-0025-4d13-a17c-a6993cb44eda-kube-api-access-4nhxw\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.152899 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-svc\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.152953 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-config-data\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.152994 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-config-data\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153052 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153095 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-595mb\" (UniqueName: \"kubernetes.io/projected/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-kube-api-access-595mb\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153121 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-config-data\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153157 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153221 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-logs\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153262 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-config\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153291 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfs9w\" (UniqueName: \"kubernetes.io/projected/02582ac3-7293-4d98-9671-f1d68ad01013-kube-api-access-qfs9w\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153352 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz569\" (UniqueName: \"kubernetes.io/projected/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-kube-api-access-pz569\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153382 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02582ac3-7293-4d98-9671-f1d68ad01013-logs\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153414 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153438 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153473 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-logs\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153524 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfv55\" (UniqueName: \"kubernetes.io/projected/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-kube-api-access-mfv55\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.153554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.154840 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-config\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.154948 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-svc\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.155145 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.155532 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.158470 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.159668 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-credential-keys\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.160381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-fernet-keys\") pod \"keystone-bootstrap-z5v4c\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.189967 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfv55\" (UniqueName: \"kubernetes.io/projected/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-kube-api-access-mfv55\") pod \"dnsmasq-dns-5f5f8c579-4ls4k\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.258818 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-48e5-account-create-6klhk"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.262241 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263816 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-logs\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263847 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfs9w\" (UniqueName: \"kubernetes.io/projected/02582ac3-7293-4d98-9671-f1d68ad01013-kube-api-access-qfs9w\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263887 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz569\" (UniqueName: \"kubernetes.io/projected/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-kube-api-access-pz569\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263913 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02582ac3-7293-4d98-9671-f1d68ad01013-logs\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-logs\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.263990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.264005 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-config-data\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.264026 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-config-data\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.264074 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-595mb\" (UniqueName: \"kubernetes.io/projected/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-kube-api-access-595mb\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.264093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-config-data\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.264118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.279856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.280551 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.281079 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-48e5-account-create-6klhk"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.283219 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.283570 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-logs\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.286245 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.286504 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-logs\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.286670 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.287535 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c59868cf7-9vlkn"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.292390 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02582ac3-7293-4d98-9671-f1d68ad01013-logs\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.299899 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.304834 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.310793 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.311742 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.311910 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-8gxg5" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.322879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfs9w\" (UniqueName: \"kubernetes.io/projected/02582ac3-7293-4d98-9671-f1d68ad01013-kube-api-access-qfs9w\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.329605 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c59868cf7-9vlkn"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.334367 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-config-data\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.334660 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.339042 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-595mb\" (UniqueName: \"kubernetes.io/projected/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-kube-api-access-595mb\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.340222 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-config-data\") pod \"watcher-applier-0\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.334380 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-config-data\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.345790 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz569\" (UniqueName: \"kubernetes.io/projected/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-kube-api-access-pz569\") pod \"watcher-decision-engine-0\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.353053 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.355934 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.360045 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.360320 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.365343 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08e6d63d-71ba-44d2-916f-ed3dd20ce845-horizon-secret-key\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.365417 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6zjd\" (UniqueName: \"kubernetes.io/projected/08e6d63d-71ba-44d2-916f-ed3dd20ce845-kube-api-access-s6zjd\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.365437 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-config-data\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.365459 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08e6d63d-71ba-44d2-916f-ed3dd20ce845-logs\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.365526 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-scripts\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.365542 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq42d\" (UniqueName: \"kubernetes.io/projected/6e8a5a7f-0d0d-4213-9813-611570f3a740-kube-api-access-rq42d\") pod \"cinder-48e5-account-create-6klhk\" (UID: \"6e8a5a7f-0d0d-4213-9813-611570f3a740\") " pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.395125 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.405458 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.440272 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gvgtl"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.441722 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.446936 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.451703 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.451876 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x7b2f" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.451959 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.473569 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5f8c579-4ls4k"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08e6d63d-71ba-44d2-916f-ed3dd20ce845-horizon-secret-key\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474718 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6zjd\" (UniqueName: \"kubernetes.io/projected/08e6d63d-71ba-44d2-916f-ed3dd20ce845-kube-api-access-s6zjd\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474744 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-config-data\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474781 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08e6d63d-71ba-44d2-916f-ed3dd20ce845-logs\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474844 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-log-httpd\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474898 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-config-data\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474968 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-scripts\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.474990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq42d\" (UniqueName: \"kubernetes.io/projected/6e8a5a7f-0d0d-4213-9813-611570f3a740-kube-api-access-rq42d\") pod \"cinder-48e5-account-create-6klhk\" (UID: \"6e8a5a7f-0d0d-4213-9813-611570f3a740\") " pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.475009 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ql97\" (UniqueName: \"kubernetes.io/projected/f3476db5-0c85-4b2c-9378-ac237c3766fa-kube-api-access-7ql97\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.475034 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.475056 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-run-httpd\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.475071 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-scripts\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.475102 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.475134 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.476289 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-scripts\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.477797 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08e6d63d-71ba-44d2-916f-ed3dd20ce845-logs\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.479180 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-config-data\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.486650 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b9df8697c-5lvvt"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.488824 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.489845 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08e6d63d-71ba-44d2-916f-ed3dd20ce845-horizon-secret-key\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.503456 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq42d\" (UniqueName: \"kubernetes.io/projected/6e8a5a7f-0d0d-4213-9813-611570f3a740-kube-api-access-rq42d\") pod \"cinder-48e5-account-create-6klhk\" (UID: \"6e8a5a7f-0d0d-4213-9813-611570f3a740\") " pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.503951 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gvgtl"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.513234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6zjd\" (UniqueName: \"kubernetes.io/projected/08e6d63d-71ba-44d2-916f-ed3dd20ce845-kube-api-access-s6zjd\") pod \"horizon-c59868cf7-9vlkn\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.549912 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b9df8697c-5lvvt"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.568754 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577163 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-log-httpd\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-config-data\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577286 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef793c0b-f866-4988-b559-e98432d41993-logs\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577313 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-config-data\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577364 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ql97\" (UniqueName: \"kubernetes.io/projected/f3476db5-0c85-4b2c-9378-ac237c3766fa-kube-api-access-7ql97\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577391 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577415 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-run-httpd\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577431 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-scripts\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577468 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577540 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr2hc\" (UniqueName: \"kubernetes.io/projected/ef793c0b-f866-4988-b559-e98432d41993-kube-api-access-dr2hc\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-config-data\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577581 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-scripts\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577622 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsld9\" (UniqueName: \"kubernetes.io/projected/bf7199a1-5e63-4d53-93dd-5818fb4e0454-kube-api-access-tsld9\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577641 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf7199a1-5e63-4d53-93dd-5818fb4e0454-logs\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-scripts\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577684 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef793c0b-f866-4988-b559-e98432d41993-horizon-secret-key\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.577702 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-combined-ca-bundle\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.578165 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.578512 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-log-httpd\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.579932 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-run-httpd\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.581431 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bb05-account-create-6czld"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.584677 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.585832 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.588877 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bb05-account-create-6czld"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.590746 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.597415 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerStarted","Data":"cef19952d2ea95e04fd898fa6be4266495564473841e4e2300f410ecd31b8aa0"} Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.621801 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-config-data\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.622136 4926 generic.go:334] "Generic (PLEG): container finished" podID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerID="82dd6763f4f094f3425331e1a229a95a8393fc34426d5d0ee8dc099bf7747414" exitCode=0 Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.622170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" event={"ID":"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6","Type":"ContainerDied","Data":"82dd6763f4f094f3425331e1a229a95a8393fc34426d5d0ee8dc099bf7747414"} Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.641225 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-scripts\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.654844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.667863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ql97\" (UniqueName: \"kubernetes.io/projected/f3476db5-0c85-4b2c-9378-ac237c3766fa-kube-api-access-7ql97\") pod \"ceilometer-0\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691277 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr2hc\" (UniqueName: \"kubernetes.io/projected/ef793c0b-f866-4988-b559-e98432d41993-kube-api-access-dr2hc\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691335 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-config-data\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691352 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-scripts\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsld9\" (UniqueName: \"kubernetes.io/projected/bf7199a1-5e63-4d53-93dd-5818fb4e0454-kube-api-access-tsld9\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691398 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf7199a1-5e63-4d53-93dd-5818fb4e0454-logs\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691418 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-scripts\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691448 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef793c0b-f866-4988-b559-e98432d41993-horizon-secret-key\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691475 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-combined-ca-bundle\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691522 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef793c0b-f866-4988-b559-e98432d41993-logs\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691540 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-config-data\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.691610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8mw8\" (UniqueName: \"kubernetes.io/projected/806823df-1d76-4cb9-a93b-4af0cbd135e3-kube-api-access-t8mw8\") pod \"neutron-bb05-account-create-6czld\" (UID: \"806823df-1d76-4cb9-a93b-4af0cbd135e3\") " pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.692892 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-config-data\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.694455 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef793c0b-f866-4988-b559-e98432d41993-logs\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.696302 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf7199a1-5e63-4d53-93dd-5818fb4e0454-logs\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.696979 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-scripts\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.727890 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsld9\" (UniqueName: \"kubernetes.io/projected/bf7199a1-5e63-4d53-93dd-5818fb4e0454-kube-api-access-tsld9\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.728108 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.748598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr2hc\" (UniqueName: \"kubernetes.io/projected/ef793c0b-f866-4988-b559-e98432d41993-kube-api-access-dr2hc\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.756111 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef793c0b-f866-4988-b559-e98432d41993-horizon-secret-key\") pod \"horizon-5b9df8697c-5lvvt\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.756391 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-scripts\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.756664 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.763948 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-config-data\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.771165 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-combined-ca-bundle\") pod \"placement-db-sync-gvgtl\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.777619 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.793419 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8mw8\" (UniqueName: \"kubernetes.io/projected/806823df-1d76-4cb9-a93b-4af0cbd135e3-kube-api-access-t8mw8\") pod \"neutron-bb05-account-create-6czld\" (UID: \"806823df-1d76-4cb9-a93b-4af0cbd135e3\") " pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.804158 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.848978 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.873055 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8mw8\" (UniqueName: \"kubernetes.io/projected/806823df-1d76-4cb9-a93b-4af0cbd135e3-kube-api-access-t8mw8\") pod \"neutron-bb05-account-create-6czld\" (UID: \"806823df-1d76-4cb9-a93b-4af0cbd135e3\") " pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.946503 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.970852 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-544499d945-t6gsp"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.975105 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-544499d945-t6gsp"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.975145 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:24 crc kubenswrapper[4926]: I1007 21:13:24.984480 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.029054 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2sj9k"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.036451 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.036717 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2sj9k"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.036794 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.036832 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.045733 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.058975 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.063103 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.063302 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.064890 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lnm8t" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.069392 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.072014 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.079659 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-96wbv" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.080723 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.082674 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.082992 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105452 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105504 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s485t\" (UniqueName: \"kubernetes.io/projected/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-kube-api-access-s485t\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105568 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105599 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-db-sync-config-data\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105624 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-svc\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105643 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105682 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-swift-storage-0\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpgjw\" (UniqueName: \"kubernetes.io/projected/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-kube-api-access-zpgjw\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105741 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-combined-ca-bundle\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105761 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-nb\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105790 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105817 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r9hm\" (UniqueName: \"kubernetes.io/projected/efa18a6d-e823-4b88-9e0c-30e0361f6b98-kube-api-access-8r9hm\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105856 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-logs\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-config\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.105925 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-sb\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.213405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-logs\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.213773 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-config\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.213855 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-sb\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.213931 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214021 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s485t\" (UniqueName: \"kubernetes.io/projected/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-kube-api-access-s485t\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214224 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214256 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-db-sync-config-data\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214299 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-svc\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214325 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-swift-storage-0\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpgjw\" (UniqueName: \"kubernetes.io/projected/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-kube-api-access-zpgjw\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214487 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-combined-ca-bundle\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214564 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-nb\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r9hm\" (UniqueName: \"kubernetes.io/projected/efa18a6d-e823-4b88-9e0c-30e0361f6b98-kube-api-access-8r9hm\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.214940 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.216810 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-logs\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.222462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-svc\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.223044 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-nb\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.225110 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-sb\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.228581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-db-sync-config-data\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.235285 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-swift-storage-0\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.235802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-config\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.238030 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.244306 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.247339 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.247817 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpgjw\" (UniqueName: \"kubernetes.io/projected/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-kube-api-access-zpgjw\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.252619 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s485t\" (UniqueName: \"kubernetes.io/projected/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-kube-api-access-s485t\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.254001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-combined-ca-bundle\") pod \"barbican-db-sync-2sj9k\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.259496 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.265102 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.273985 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r9hm\" (UniqueName: \"kubernetes.io/projected/efa18a6d-e823-4b88-9e0c-30e0361f6b98-kube-api-access-8r9hm\") pod \"dnsmasq-dns-544499d945-t6gsp\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.316907 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324185 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324248 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324311 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-logs\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324348 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429wm\" (UniqueName: \"kubernetes.io/projected/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-kube-api-access-429wm\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324381 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324425 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324447 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.324484 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.340406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.427820 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-logs\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.427888 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429wm\" (UniqueName: \"kubernetes.io/projected/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-kube-api-access-429wm\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.427915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.427962 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.427985 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.428020 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.428066 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.428081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.428793 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-logs\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.429038 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.433461 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.433580 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.435498 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.443539 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.449776 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429wm\" (UniqueName: \"kubernetes.io/projected/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-kube-api-access-429wm\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.453500 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.473968 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.479753 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.515134 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.515753 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.644998 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-z5v4c"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.656924 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.687113 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.734833 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-config\") pod \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.734934 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-sb\") pod \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.735002 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-nb\") pod \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.735052 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-swift-storage-0\") pod \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.735136 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-svc\") pod \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.735238 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpqgq\" (UniqueName: \"kubernetes.io/projected/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-kube-api-access-cpqgq\") pod \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\" (UID: \"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6\") " Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.754420 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-kube-api-access-cpqgq" (OuterVolumeSpecName: "kube-api-access-cpqgq") pod "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" (UID: "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6"). InnerVolumeSpecName "kube-api-access-cpqgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.826467 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-config" (OuterVolumeSpecName: "config") pod "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" (UID: "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.841941 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" (UID: "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.845436 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.846992 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" (UID: "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.847898 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.847924 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpqgq\" (UniqueName: \"kubernetes.io/projected/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-kube-api-access-cpqgq\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.850785 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" (UID: "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.859041 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" (UID: "2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.952154 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.952608 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:25 crc kubenswrapper[4926]: I1007 21:13:25.952637 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.124093 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gvgtl"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.144685 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bb05-account-create-6czld"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.155684 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5f8c579-4ls4k"] Oct 07 21:13:26 crc kubenswrapper[4926]: W1007 21:13:26.158969 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dea945e_c5cb_4d6f_b0df_fb2207cb0b89.slice/crio-1b656de4172b5067eabeb4e44b83d88d0d124ed413d4a930ee82196981cb9f13 WatchSource:0}: Error finding container 1b656de4172b5067eabeb4e44b83d88d0d124ed413d4a930ee82196981cb9f13: Status 404 returned error can't find the container with id 1b656de4172b5067eabeb4e44b83d88d0d124ed413d4a930ee82196981cb9f13 Oct 07 21:13:26 crc kubenswrapper[4926]: W1007 21:13:26.161389 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod806823df_1d76_4cb9_a93b_4af0cbd135e3.slice/crio-9d5aa67a45e95a0f4e508a0c37fecc4971fb4edca25b1551ca10da6bd158b198 WatchSource:0}: Error finding container 9d5aa67a45e95a0f4e508a0c37fecc4971fb4edca25b1551ca10da6bd158b198: Status 404 returned error can't find the container with id 9d5aa67a45e95a0f4e508a0c37fecc4971fb4edca25b1551ca10da6bd158b198 Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.165447 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.440764 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-48e5-account-create-6klhk"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.465974 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b9df8697c-5lvvt"] Oct 07 21:13:26 crc kubenswrapper[4926]: W1007 21:13:26.467462 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef793c0b_f866_4988_b559_e98432d41993.slice/crio-952f6efcafc5500afb5315f269a2089b510f4199d66f4f35ad90c9522fc959d9 WatchSource:0}: Error finding container 952f6efcafc5500afb5315f269a2089b510f4199d66f4f35ad90c9522fc959d9: Status 404 returned error can't find the container with id 952f6efcafc5500afb5315f269a2089b510f4199d66f4f35ad90c9522fc959d9 Oct 07 21:13:26 crc kubenswrapper[4926]: W1007 21:13:26.485212 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e8a5a7f_0d0d_4213_9813_611570f3a740.slice/crio-92098bb510c4166896c64fcdf074a769a700e9f088bbd929d7dc0410151017a5 WatchSource:0}: Error finding container 92098bb510c4166896c64fcdf074a769a700e9f088bbd929d7dc0410151017a5: Status 404 returned error can't find the container with id 92098bb510c4166896c64fcdf074a769a700e9f088bbd929d7dc0410151017a5 Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.487564 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c59868cf7-9vlkn"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.500924 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:13:26 crc kubenswrapper[4926]: W1007 21:13:26.507428 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbac957e_ce41_4c6b_ada3_4ebfbeb0b106.slice/crio-fd76c572cd22cfbdb51e46260814eba6f58b7aa9acd9e380cab294dfcc6947b1 WatchSource:0}: Error finding container fd76c572cd22cfbdb51e46260814eba6f58b7aa9acd9e380cab294dfcc6947b1: Status 404 returned error can't find the container with id fd76c572cd22cfbdb51e46260814eba6f58b7aa9acd9e380cab294dfcc6947b1 Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.524972 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.663362 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"02582ac3-7293-4d98-9671-f1d68ad01013","Type":"ContainerStarted","Data":"e8b69768c1a203c24195549018320d7b514cbec5bcdbd709dfd5fb508e3fbba4"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.663425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"02582ac3-7293-4d98-9671-f1d68ad01013","Type":"ContainerStarted","Data":"b5917365b23b7add7c4d569c40fd0fed548587c32043a71d30edf53797ff1b48"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.663437 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"02582ac3-7293-4d98-9671-f1d68ad01013","Type":"ContainerStarted","Data":"684ddc8d22b9838b53dcaaaf1128037e88dd5af5d9a9004c07f3237607f4fb53"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.664033 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.665822 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89","Type":"ContainerStarted","Data":"1b656de4172b5067eabeb4e44b83d88d0d124ed413d4a930ee82196981cb9f13"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.669447 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": dial tcp 10.217.0.149:9322: connect: connection refused" Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.674803 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-48e5-account-create-6klhk" event={"ID":"6e8a5a7f-0d0d-4213-9813-611570f3a740","Type":"ContainerStarted","Data":"92098bb510c4166896c64fcdf074a769a700e9f088bbd929d7dc0410151017a5"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.694111 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=3.694082411 podStartE2EDuration="3.694082411s" podCreationTimestamp="2025-10-07 21:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:26.687458749 +0000 UTC m=+1056.726039939" watchObservedRunningTime="2025-10-07 21:13:26.694082411 +0000 UTC m=+1056.732663551" Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.718795 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106","Type":"ContainerStarted","Data":"fd76c572cd22cfbdb51e46260814eba6f58b7aa9acd9e380cab294dfcc6947b1"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.721266 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b9df8697c-5lvvt" event={"ID":"ef793c0b-f866-4988-b559-e98432d41993","Type":"ContainerStarted","Data":"952f6efcafc5500afb5315f269a2089b510f4199d66f4f35ad90c9522fc959d9"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.726789 4926 generic.go:334] "Generic (PLEG): container finished" podID="2150cef6-8b1a-4f79-8bc9-fe1f772bf319" containerID="2e9128e3bdb80365400275ff6467954808b7b5a2d9d788ff41700c0b4136bc18" exitCode=0 Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.726843 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" event={"ID":"2150cef6-8b1a-4f79-8bc9-fe1f772bf319","Type":"ContainerDied","Data":"2e9128e3bdb80365400275ff6467954808b7b5a2d9d788ff41700c0b4136bc18"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.726869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" event={"ID":"2150cef6-8b1a-4f79-8bc9-fe1f772bf319","Type":"ContainerStarted","Data":"bf1c5b646d39c9a78072c6a51d665847cad0b4f217a0ef2afc43a6dcd3656beb"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.733236 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z5v4c" event={"ID":"920e2024-0025-4d13-a17c-a6993cb44eda","Type":"ContainerStarted","Data":"09e6b54c7a805f24299659965dc167ee53ce559d4da466c095790298d53440d5"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.733286 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z5v4c" event={"ID":"920e2024-0025-4d13-a17c-a6993cb44eda","Type":"ContainerStarted","Data":"7e327bd505eed6d5c08d50a818a628b3c60332a6a084ebd1494ff97577e676be"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.744831 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3476db5-0c85-4b2c-9378-ac237c3766fa","Type":"ContainerStarted","Data":"d010e8d17d60781e94e6e378d9cfe69ec2dc204a92406284bc6afbe9755f4298"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.809076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c59868cf7-9vlkn" event={"ID":"08e6d63d-71ba-44d2-916f-ed3dd20ce845","Type":"ContainerStarted","Data":"5abd251844d79c7c55563e6936e7a69cd989aad94b56f10237a213c937eb94e9"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.815173 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvgtl" event={"ID":"bf7199a1-5e63-4d53-93dd-5818fb4e0454","Type":"ContainerStarted","Data":"48bc7a3df5741c1c8edf8fa12deb8624247826e2324b62ff2b4a79fdd88950a9"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.829281 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-544499d945-t6gsp"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.832974 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-z5v4c" podStartSLOduration=3.8329547010000002 podStartE2EDuration="3.832954701s" podCreationTimestamp="2025-10-07 21:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:26.805954567 +0000 UTC m=+1056.844535717" watchObservedRunningTime="2025-10-07 21:13:26.832954701 +0000 UTC m=+1056.871535851" Oct 07 21:13:26 crc kubenswrapper[4926]: W1007 21:13:26.856418 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bab0bbe_8fa5_451f_89fe_a22426ecdfe2.slice/crio-1e130eea95b2325a617ffa9fa5d752bdd74d577badba1dd35b22b07452e6dc29 WatchSource:0}: Error finding container 1e130eea95b2325a617ffa9fa5d752bdd74d577badba1dd35b22b07452e6dc29: Status 404 returned error can't find the container with id 1e130eea95b2325a617ffa9fa5d752bdd74d577badba1dd35b22b07452e6dc29 Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.896293 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.896928 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db7bcf759-7zm8s" event={"ID":"2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6","Type":"ContainerDied","Data":"6da72074707ec2106eeccd49fe846e3c0da7176cafffeafbf08b76c1ef77a84c"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.897006 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2sj9k"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.897033 4926 scope.go:117] "RemoveContainer" containerID="82dd6763f4f094f3425331e1a229a95a8393fc34426d5d0ee8dc099bf7747414" Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.941334 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bb05-account-create-6czld" event={"ID":"806823df-1d76-4cb9-a93b-4af0cbd135e3","Type":"ContainerStarted","Data":"3f38ef18fd93b1a2ec80e15808b80341c213da5eda9506eaea8033dc03225601"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.941372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bb05-account-create-6czld" event={"ID":"806823df-1d76-4cb9-a93b-4af0cbd135e3","Type":"ContainerStarted","Data":"9d5aa67a45e95a0f4e508a0c37fecc4971fb4edca25b1551ca10da6bd158b198"} Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.965475 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db7bcf759-7zm8s"] Oct 07 21:13:26 crc kubenswrapper[4926]: I1007 21:13:26.978813 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5db7bcf759-7zm8s"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.011937 4926 scope.go:117] "RemoveContainer" containerID="9fdff9bfbc66f22cd924bbc3adc3ee8f4b86de87aad4ffe11be6ab776a44d092" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.018886 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.560957 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.616272 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b9df8697c-5lvvt"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.641776 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.654248 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-dd68f6c57-k5dpz"] Oct 07 21:13:27 crc kubenswrapper[4926]: E1007 21:13:27.654778 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerName="dnsmasq-dns" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.654801 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerName="dnsmasq-dns" Oct 07 21:13:27 crc kubenswrapper[4926]: E1007 21:13:27.654842 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerName="init" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.654850 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerName="init" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.655064 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" containerName="dnsmasq-dns" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.656621 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.663750 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dd68f6c57-k5dpz"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.683127 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.779139 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.788412 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.811497 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.824076 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eff2499a-3993-4955-b32c-86883e6d649f-logs\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.824150 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-scripts\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.824173 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnmch\" (UniqueName: \"kubernetes.io/projected/eff2499a-3993-4955-b32c-86883e6d649f-kube-api-access-mnmch\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.824249 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eff2499a-3993-4955-b32c-86883e6d649f-horizon-secret-key\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.824304 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-config-data\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: E1007 21:13:27.838694 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e8a5a7f_0d0d_4213_9813_611570f3a740.slice/crio-518fbd721f2fdbaddfe0ff3da576094ecf27c602f6ba9ffb9d4dc2839caf860f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e8a5a7f_0d0d_4213_9813_611570f3a740.slice/crio-conmon-518fbd721f2fdbaddfe0ff3da576094ecf27c602f6ba9ffb9d4dc2839caf860f.scope\": RecentStats: unable to find data in memory cache]" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926421 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-nb\") pod \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926493 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-svc\") pod \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926560 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfv55\" (UniqueName: \"kubernetes.io/projected/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-kube-api-access-mfv55\") pod \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926651 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-swift-storage-0\") pod \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926689 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-sb\") pod \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926705 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-config\") pod \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\" (UID: \"2150cef6-8b1a-4f79-8bc9-fe1f772bf319\") " Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.926989 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eff2499a-3993-4955-b32c-86883e6d649f-logs\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.927038 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-scripts\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.927065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnmch\" (UniqueName: \"kubernetes.io/projected/eff2499a-3993-4955-b32c-86883e6d649f-kube-api-access-mnmch\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.927102 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eff2499a-3993-4955-b32c-86883e6d649f-horizon-secret-key\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.927144 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-config-data\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.928381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-config-data\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.928683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eff2499a-3993-4955-b32c-86883e6d649f-logs\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.929172 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-scripts\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.934669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eff2499a-3993-4955-b32c-86883e6d649f-horizon-secret-key\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.939738 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-kube-api-access-mfv55" (OuterVolumeSpecName: "kube-api-access-mfv55") pod "2150cef6-8b1a-4f79-8bc9-fe1f772bf319" (UID: "2150cef6-8b1a-4f79-8bc9-fe1f772bf319"). InnerVolumeSpecName "kube-api-access-mfv55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.953078 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnmch\" (UniqueName: \"kubernetes.io/projected/eff2499a-3993-4955-b32c-86883e6d649f-kube-api-access-mnmch\") pod \"horizon-dd68f6c57-k5dpz\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.986747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2150cef6-8b1a-4f79-8bc9-fe1f772bf319" (UID: "2150cef6-8b1a-4f79-8bc9-fe1f772bf319"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:27 crc kubenswrapper[4926]: I1007 21:13:27.989709 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2150cef6-8b1a-4f79-8bc9-fe1f772bf319" (UID: "2150cef6-8b1a-4f79-8bc9-fe1f772bf319"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.000988 4926 generic.go:334] "Generic (PLEG): container finished" podID="6e8a5a7f-0d0d-4213-9813-611570f3a740" containerID="518fbd721f2fdbaddfe0ff3da576094ecf27c602f6ba9ffb9d4dc2839caf860f" exitCode=0 Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.001155 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-48e5-account-create-6klhk" event={"ID":"6e8a5a7f-0d0d-4213-9813-611570f3a740","Type":"ContainerDied","Data":"518fbd721f2fdbaddfe0ff3da576094ecf27c602f6ba9ffb9d4dc2839caf860f"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.001339 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-config" (OuterVolumeSpecName: "config") pod "2150cef6-8b1a-4f79-8bc9-fe1f772bf319" (UID: "2150cef6-8b1a-4f79-8bc9-fe1f772bf319"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.008943 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2150cef6-8b1a-4f79-8bc9-fe1f772bf319" (UID: "2150cef6-8b1a-4f79-8bc9-fe1f772bf319"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.011811 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2150cef6-8b1a-4f79-8bc9-fe1f772bf319" (UID: "2150cef6-8b1a-4f79-8bc9-fe1f772bf319"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.029461 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.029493 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.029505 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfv55\" (UniqueName: \"kubernetes.io/projected/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-kube-api-access-mfv55\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.029517 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.029528 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.029536 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150cef6-8b1a-4f79-8bc9-fe1f772bf319-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.030129 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5f61f403-e0b6-45f1-a0a8-d073f1bdd190","Type":"ContainerStarted","Data":"e8585ba81c74f64cd5b2c8063e4a13c43b140638db05e2af25f4cdc350f21a3a"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.036478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2sj9k" event={"ID":"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2","Type":"ContainerStarted","Data":"1e130eea95b2325a617ffa9fa5d752bdd74d577badba1dd35b22b07452e6dc29"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.051566 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.051772 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5f8c579-4ls4k" event={"ID":"2150cef6-8b1a-4f79-8bc9-fe1f772bf319","Type":"ContainerDied","Data":"bf1c5b646d39c9a78072c6a51d665847cad0b4f217a0ef2afc43a6dcd3656beb"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.051894 4926 scope.go:117] "RemoveContainer" containerID="2e9128e3bdb80365400275ff6467954808b7b5a2d9d788ff41700c0b4136bc18" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.055529 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9","Type":"ContainerStarted","Data":"830d5a7dfaee87cc9978fe8830708b379543b7398ac99a1ddd4ff908c59a4265"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.078938 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544499d945-t6gsp" event={"ID":"efa18a6d-e823-4b88-9e0c-30e0361f6b98","Type":"ContainerStarted","Data":"4701694df8bf30c6a8c6493b44b997b38a6d956cf0b6c5f94d3cdc9f6e116d9e"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.079003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544499d945-t6gsp" event={"ID":"efa18a6d-e823-4b88-9e0c-30e0361f6b98","Type":"ContainerStarted","Data":"db74c6e0c10427cdbd6cf117d43966a96bbed0fcb713a078492a2b14d1a0c9ba"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.094889 4926 generic.go:334] "Generic (PLEG): container finished" podID="806823df-1d76-4cb9-a93b-4af0cbd135e3" containerID="3f38ef18fd93b1a2ec80e15808b80341c213da5eda9506eaea8033dc03225601" exitCode=0 Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.094992 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bb05-account-create-6czld" event={"ID":"806823df-1d76-4cb9-a93b-4af0cbd135e3","Type":"ContainerDied","Data":"3f38ef18fd93b1a2ec80e15808b80341c213da5eda9506eaea8033dc03225601"} Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.103443 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.437934 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5f8c579-4ls4k"] Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.456329 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f5f8c579-4ls4k"] Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.698682 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2150cef6-8b1a-4f79-8bc9-fe1f772bf319" path="/var/lib/kubelet/pods/2150cef6-8b1a-4f79-8bc9-fe1f772bf319/volumes" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.709301 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6" path="/var/lib/kubelet/pods/2422ef1e-9a51-4d5e-b1ed-56bccf11fbd6/volumes" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.774414 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.944814 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dd68f6c57-k5dpz"] Oct 07 21:13:28 crc kubenswrapper[4926]: I1007 21:13:28.974660 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8mw8\" (UniqueName: \"kubernetes.io/projected/806823df-1d76-4cb9-a93b-4af0cbd135e3-kube-api-access-t8mw8\") pod \"806823df-1d76-4cb9-a93b-4af0cbd135e3\" (UID: \"806823df-1d76-4cb9-a93b-4af0cbd135e3\") " Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.001437 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/806823df-1d76-4cb9-a93b-4af0cbd135e3-kube-api-access-t8mw8" (OuterVolumeSpecName: "kube-api-access-t8mw8") pod "806823df-1d76-4cb9-a93b-4af0cbd135e3" (UID: "806823df-1d76-4cb9-a93b-4af0cbd135e3"). InnerVolumeSpecName "kube-api-access-t8mw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.078460 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8mw8\" (UniqueName: \"kubernetes.io/projected/806823df-1d76-4cb9-a93b-4af0cbd135e3-kube-api-access-t8mw8\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.125491 4926 generic.go:334] "Generic (PLEG): container finished" podID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerID="4701694df8bf30c6a8c6493b44b997b38a6d956cf0b6c5f94d3cdc9f6e116d9e" exitCode=0 Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.125561 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544499d945-t6gsp" event={"ID":"efa18a6d-e823-4b88-9e0c-30e0361f6b98","Type":"ContainerDied","Data":"4701694df8bf30c6a8c6493b44b997b38a6d956cf0b6c5f94d3cdc9f6e116d9e"} Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.125588 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544499d945-t6gsp" event={"ID":"efa18a6d-e823-4b88-9e0c-30e0361f6b98","Type":"ContainerStarted","Data":"f09c937d588e21c7a130574801d8780b483960c6786825e920895c89cd6b25c0"} Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.126632 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.128429 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bb05-account-create-6czld" event={"ID":"806823df-1d76-4cb9-a93b-4af0cbd135e3","Type":"ContainerDied","Data":"9d5aa67a45e95a0f4e508a0c37fecc4971fb4edca25b1551ca10da6bd158b198"} Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.128479 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api-log" containerID="cri-o://b5917365b23b7add7c4d569c40fd0fed548587c32043a71d30edf53797ff1b48" gracePeriod=30 Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.128507 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d5aa67a45e95a0f4e508a0c37fecc4971fb4edca25b1551ca10da6bd158b198" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.128648 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" containerID="cri-o://e8b69768c1a203c24195549018320d7b514cbec5bcdbd709dfd5fb508e3fbba4" gracePeriod=30 Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.131359 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bb05-account-create-6czld" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.132580 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": EOF" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.135761 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": EOF" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.151896 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-544499d945-t6gsp" podStartSLOduration=5.151870461 podStartE2EDuration="5.151870461s" podCreationTimestamp="2025-10-07 21:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:29.144752255 +0000 UTC m=+1059.183333405" watchObservedRunningTime="2025-10-07 21:13:29.151870461 +0000 UTC m=+1059.190451611" Oct 07 21:13:29 crc kubenswrapper[4926]: I1007 21:13:29.406750 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 07 21:13:30 crc kubenswrapper[4926]: I1007 21:13:30.145588 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9","Type":"ContainerStarted","Data":"a13fa194e03bde11aa16a3ab57643b48bfbcb49b3c506611dc887dbab96cc31f"} Oct 07 21:13:30 crc kubenswrapper[4926]: I1007 21:13:30.148183 4926 generic.go:334] "Generic (PLEG): container finished" podID="02582ac3-7293-4d98-9671-f1d68ad01013" containerID="b5917365b23b7add7c4d569c40fd0fed548587c32043a71d30edf53797ff1b48" exitCode=143 Oct 07 21:13:30 crc kubenswrapper[4926]: I1007 21:13:30.148251 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"02582ac3-7293-4d98-9671-f1d68ad01013","Type":"ContainerDied","Data":"b5917365b23b7add7c4d569c40fd0fed548587c32043a71d30edf53797ff1b48"} Oct 07 21:13:30 crc kubenswrapper[4926]: I1007 21:13:30.151112 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5f61f403-e0b6-45f1-a0a8-d073f1bdd190","Type":"ContainerStarted","Data":"bb71457dc900b962c61909e4cb2debcb69d33dd0779fb069ef52b617e460d485"} Oct 07 21:13:31 crc kubenswrapper[4926]: I1007 21:13:31.166374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dd68f6c57-k5dpz" event={"ID":"eff2499a-3993-4955-b32c-86883e6d649f","Type":"ContainerStarted","Data":"1310059a53f84b291395ec24cd342a7492f32fbf2274671933fef56c30f36df3"} Oct 07 21:13:31 crc kubenswrapper[4926]: I1007 21:13:31.877432 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": read tcp 10.217.0.2:42750->10.217.0.149:9322: read: connection reset by peer" Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.199261 4926 generic.go:334] "Generic (PLEG): container finished" podID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerID="cef19952d2ea95e04fd898fa6be4266495564473841e4e2300f410ecd31b8aa0" exitCode=0 Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.199351 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerDied","Data":"cef19952d2ea95e04fd898fa6be4266495564473841e4e2300f410ecd31b8aa0"} Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.203105 4926 generic.go:334] "Generic (PLEG): container finished" podID="920e2024-0025-4d13-a17c-a6993cb44eda" containerID="09e6b54c7a805f24299659965dc167ee53ce559d4da466c095790298d53440d5" exitCode=0 Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.203235 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z5v4c" event={"ID":"920e2024-0025-4d13-a17c-a6993cb44eda","Type":"ContainerDied","Data":"09e6b54c7a805f24299659965dc167ee53ce559d4da466c095790298d53440d5"} Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.205240 4926 generic.go:334] "Generic (PLEG): container finished" podID="02582ac3-7293-4d98-9671-f1d68ad01013" containerID="e8b69768c1a203c24195549018320d7b514cbec5bcdbd709dfd5fb508e3fbba4" exitCode=0 Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.205277 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"02582ac3-7293-4d98-9671-f1d68ad01013","Type":"ContainerDied","Data":"e8b69768c1a203c24195549018320d7b514cbec5bcdbd709dfd5fb508e3fbba4"} Oct 07 21:13:32 crc kubenswrapper[4926]: I1007 21:13:32.988569 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c59868cf7-9vlkn"] Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.029139 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bf67dc754-22mcb"] Oct 07 21:13:33 crc kubenswrapper[4926]: E1007 21:13:33.030469 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="806823df-1d76-4cb9-a93b-4af0cbd135e3" containerName="mariadb-account-create" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.030497 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="806823df-1d76-4cb9-a93b-4af0cbd135e3" containerName="mariadb-account-create" Oct 07 21:13:33 crc kubenswrapper[4926]: E1007 21:13:33.030527 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2150cef6-8b1a-4f79-8bc9-fe1f772bf319" containerName="init" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.030535 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2150cef6-8b1a-4f79-8bc9-fe1f772bf319" containerName="init" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.030747 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="806823df-1d76-4cb9-a93b-4af0cbd135e3" containerName="mariadb-account-create" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.030769 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2150cef6-8b1a-4f79-8bc9-fe1f772bf319" containerName="init" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.032788 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.035847 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.040527 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bf67dc754-22mcb"] Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.093308 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-dd68f6c57-k5dpz"] Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.116979 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-858b4c475d-fcp7q"] Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.118503 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.151669 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-858b4c475d-fcp7q"] Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.171479 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-tls-certs\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.171534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-combined-ca-bundle\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.171898 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npwj9\" (UniqueName: \"kubernetes.io/projected/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-kube-api-access-npwj9\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.171954 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-scripts\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.172155 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-logs\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.172217 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-config-data\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.172319 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-secret-key\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274275 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-tls-certs\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274561 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-combined-ca-bundle\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274588 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42be0d4-2ced-4d32-84f4-49389e2a0e82-logs\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274607 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-combined-ca-bundle\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274626 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mrsj\" (UniqueName: \"kubernetes.io/projected/c42be0d4-2ced-4d32-84f4-49389e2a0e82-kube-api-access-6mrsj\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274656 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c42be0d4-2ced-4d32-84f4-49389e2a0e82-scripts\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274805 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npwj9\" (UniqueName: \"kubernetes.io/projected/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-kube-api-access-npwj9\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274908 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-horizon-secret-key\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.274945 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-scripts\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.275067 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-logs\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.275117 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-config-data\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.275153 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c42be0d4-2ced-4d32-84f4-49389e2a0e82-config-data\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.275243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-horizon-tls-certs\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.275273 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-secret-key\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.275980 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-scripts\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.276344 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-logs\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.277367 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-config-data\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.280417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-secret-key\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.282649 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-tls-certs\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.303741 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-combined-ca-bundle\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.304425 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npwj9\" (UniqueName: \"kubernetes.io/projected/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-kube-api-access-npwj9\") pod \"horizon-7bf67dc754-22mcb\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.365882 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.377259 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-combined-ca-bundle\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.377735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42be0d4-2ced-4d32-84f4-49389e2a0e82-logs\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.377828 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mrsj\" (UniqueName: \"kubernetes.io/projected/c42be0d4-2ced-4d32-84f4-49389e2a0e82-kube-api-access-6mrsj\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.377926 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c42be0d4-2ced-4d32-84f4-49389e2a0e82-scripts\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.378041 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-horizon-secret-key\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.378148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c42be0d4-2ced-4d32-84f4-49389e2a0e82-config-data\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.378048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42be0d4-2ced-4d32-84f4-49389e2a0e82-logs\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.378249 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-horizon-tls-certs\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.379275 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c42be0d4-2ced-4d32-84f4-49389e2a0e82-scripts\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.379415 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c42be0d4-2ced-4d32-84f4-49389e2a0e82-config-data\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.382060 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-horizon-secret-key\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.382268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-combined-ca-bundle\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.382412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42be0d4-2ced-4d32-84f4-49389e2a0e82-horizon-tls-certs\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.402672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mrsj\" (UniqueName: \"kubernetes.io/projected/c42be0d4-2ced-4d32-84f4-49389e2a0e82-kube-api-access-6mrsj\") pod \"horizon-858b4c475d-fcp7q\" (UID: \"c42be0d4-2ced-4d32-84f4-49389e2a0e82\") " pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.445261 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.722689 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.791911 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq42d\" (UniqueName: \"kubernetes.io/projected/6e8a5a7f-0d0d-4213-9813-611570f3a740-kube-api-access-rq42d\") pod \"6e8a5a7f-0d0d-4213-9813-611570f3a740\" (UID: \"6e8a5a7f-0d0d-4213-9813-611570f3a740\") " Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.803588 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e8a5a7f-0d0d-4213-9813-611570f3a740-kube-api-access-rq42d" (OuterVolumeSpecName: "kube-api-access-rq42d") pod "6e8a5a7f-0d0d-4213-9813-611570f3a740" (UID: "6e8a5a7f-0d0d-4213-9813-611570f3a740"). InnerVolumeSpecName "kube-api-access-rq42d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:33 crc kubenswrapper[4926]: I1007 21:13:33.893944 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq42d\" (UniqueName: \"kubernetes.io/projected/6e8a5a7f-0d0d-4213-9813-611570f3a740-kube-api-access-rq42d\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.226223 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-48e5-account-create-6klhk" event={"ID":"6e8a5a7f-0d0d-4213-9813-611570f3a740","Type":"ContainerDied","Data":"92098bb510c4166896c64fcdf074a769a700e9f088bbd929d7dc0410151017a5"} Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.226271 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92098bb510c4166896c64fcdf074a769a700e9f088bbd929d7dc0410151017a5" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.226284 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-48e5-account-create-6klhk" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.406784 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": dial tcp 10.217.0.149:9322: connect: connection refused" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.548899 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-sjjzr"] Oct 07 21:13:34 crc kubenswrapper[4926]: E1007 21:13:34.549639 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8a5a7f-0d0d-4213-9813-611570f3a740" containerName="mariadb-account-create" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.549693 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8a5a7f-0d0d-4213-9813-611570f3a740" containerName="mariadb-account-create" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.550030 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e8a5a7f-0d0d-4213-9813-611570f3a740" containerName="mariadb-account-create" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.551018 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.553506 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.554267 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.554392 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4cxdd" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.562591 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-sjjzr"] Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.715831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-combined-ca-bundle\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.716865 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-config\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.717317 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnf5g\" (UniqueName: \"kubernetes.io/projected/6aa8d106-5d6e-4585-b368-e54382bed918-kube-api-access-hnf5g\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.818919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-combined-ca-bundle\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.818999 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-config\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.819064 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnf5g\" (UniqueName: \"kubernetes.io/projected/6aa8d106-5d6e-4585-b368-e54382bed918-kube-api-access-hnf5g\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.824956 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-config\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.833139 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-combined-ca-bundle\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.836782 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnf5g\" (UniqueName: \"kubernetes.io/projected/6aa8d106-5d6e-4585-b368-e54382bed918-kube-api-access-hnf5g\") pod \"neutron-db-sync-sjjzr\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:34 crc kubenswrapper[4926]: I1007 21:13:34.880997 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:13:35 crc kubenswrapper[4926]: I1007 21:13:35.318359 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:13:35 crc kubenswrapper[4926]: I1007 21:13:35.424427 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c6866c67-6cjbf"] Oct 07 21:13:35 crc kubenswrapper[4926]: I1007 21:13:35.424701 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="dnsmasq-dns" containerID="cri-o://3ea737c649028e281a9295bbcc0e840bb716c92299249b00ab2fdfcc847219e7" gracePeriod=10 Oct 07 21:13:36 crc kubenswrapper[4926]: I1007 21:13:36.255036 4926 generic.go:334] "Generic (PLEG): container finished" podID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerID="3ea737c649028e281a9295bbcc0e840bb716c92299249b00ab2fdfcc847219e7" exitCode=0 Oct 07 21:13:36 crc kubenswrapper[4926]: I1007 21:13:36.255114 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" event={"ID":"69fa95b0-9096-4383-bfbf-3af3c095c9ad","Type":"ContainerDied","Data":"3ea737c649028e281a9295bbcc0e840bb716c92299249b00ab2fdfcc847219e7"} Oct 07 21:13:36 crc kubenswrapper[4926]: I1007 21:13:36.690466 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.136:5353: connect: connection refused" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.629102 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.804881 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-config-data\") pod \"920e2024-0025-4d13-a17c-a6993cb44eda\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.804972 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-fernet-keys\") pod \"920e2024-0025-4d13-a17c-a6993cb44eda\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.805023 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-scripts\") pod \"920e2024-0025-4d13-a17c-a6993cb44eda\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.805261 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-credential-keys\") pod \"920e2024-0025-4d13-a17c-a6993cb44eda\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.805280 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nhxw\" (UniqueName: \"kubernetes.io/projected/920e2024-0025-4d13-a17c-a6993cb44eda-kube-api-access-4nhxw\") pod \"920e2024-0025-4d13-a17c-a6993cb44eda\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.805346 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-combined-ca-bundle\") pod \"920e2024-0025-4d13-a17c-a6993cb44eda\" (UID: \"920e2024-0025-4d13-a17c-a6993cb44eda\") " Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.809547 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-scripts" (OuterVolumeSpecName: "scripts") pod "920e2024-0025-4d13-a17c-a6993cb44eda" (UID: "920e2024-0025-4d13-a17c-a6993cb44eda"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.809743 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "920e2024-0025-4d13-a17c-a6993cb44eda" (UID: "920e2024-0025-4d13-a17c-a6993cb44eda"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.812304 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920e2024-0025-4d13-a17c-a6993cb44eda-kube-api-access-4nhxw" (OuterVolumeSpecName: "kube-api-access-4nhxw") pod "920e2024-0025-4d13-a17c-a6993cb44eda" (UID: "920e2024-0025-4d13-a17c-a6993cb44eda"). InnerVolumeSpecName "kube-api-access-4nhxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.819205 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "920e2024-0025-4d13-a17c-a6993cb44eda" (UID: "920e2024-0025-4d13-a17c-a6993cb44eda"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.835672 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "920e2024-0025-4d13-a17c-a6993cb44eda" (UID: "920e2024-0025-4d13-a17c-a6993cb44eda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.837278 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-config-data" (OuterVolumeSpecName: "config-data") pod "920e2024-0025-4d13-a17c-a6993cb44eda" (UID: "920e2024-0025-4d13-a17c-a6993cb44eda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.907437 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.907468 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.907478 4926 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.907489 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nhxw\" (UniqueName: \"kubernetes.io/projected/920e2024-0025-4d13-a17c-a6993cb44eda-kube-api-access-4nhxw\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.907499 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:37 crc kubenswrapper[4926]: I1007 21:13:37.907507 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920e2024-0025-4d13-a17c-a6993cb44eda-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.278128 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z5v4c" event={"ID":"920e2024-0025-4d13-a17c-a6993cb44eda","Type":"ContainerDied","Data":"7e327bd505eed6d5c08d50a818a628b3c60332a6a084ebd1494ff97577e676be"} Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.278480 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e327bd505eed6d5c08d50a818a628b3c60332a6a084ebd1494ff97577e676be" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.278213 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z5v4c" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.784269 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-z5v4c"] Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.791575 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-z5v4c"] Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.836455 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-pvnmq"] Oct 07 21:13:38 crc kubenswrapper[4926]: E1007 21:13:38.836919 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920e2024-0025-4d13-a17c-a6993cb44eda" containerName="keystone-bootstrap" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.836932 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="920e2024-0025-4d13-a17c-a6993cb44eda" containerName="keystone-bootstrap" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.837524 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="920e2024-0025-4d13-a17c-a6993cb44eda" containerName="keystone-bootstrap" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.838267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.840217 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.841298 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-x9gv9" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.841484 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.843737 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.875247 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pvnmq"] Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.926392 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-fernet-keys\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.926466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-scripts\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.926499 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-credential-keys\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.926642 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-combined-ca-bundle\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.926758 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4rfg\" (UniqueName: \"kubernetes.io/projected/0b74a118-4677-43dd-8506-dc66795fe8a1-kube-api-access-f4rfg\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:38 crc kubenswrapper[4926]: I1007 21:13:38.926903 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-config-data\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.028222 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-fernet-keys\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.028301 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-scripts\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.028333 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-credential-keys\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.028374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-combined-ca-bundle\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.028399 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4rfg\" (UniqueName: \"kubernetes.io/projected/0b74a118-4677-43dd-8506-dc66795fe8a1-kube-api-access-f4rfg\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.028434 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-config-data\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.035298 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-fernet-keys\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.036103 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-scripts\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.042229 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-combined-ca-bundle\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.043137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-credential-keys\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.046668 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4rfg\" (UniqueName: \"kubernetes.io/projected/0b74a118-4677-43dd-8506-dc66795fe8a1-kube-api-access-f4rfg\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.050864 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-config-data\") pod \"keystone-bootstrap-pvnmq\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.165378 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.223740 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-bszbk"] Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.225844 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.228053 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hw4jl" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.228270 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.229696 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.232087 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bszbk"] Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.334071 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/920741ff-d4f2-495b-907c-d052b32c6380-etc-machine-id\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.334458 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-config-data\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.334487 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-combined-ca-bundle\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.334515 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-db-sync-config-data\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.334649 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbrwz\" (UniqueName: \"kubernetes.io/projected/920741ff-d4f2-495b-907c-d052b32c6380-kube-api-access-bbrwz\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.334683 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-scripts\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.436825 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbrwz\" (UniqueName: \"kubernetes.io/projected/920741ff-d4f2-495b-907c-d052b32c6380-kube-api-access-bbrwz\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.436896 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-scripts\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.436963 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/920741ff-d4f2-495b-907c-d052b32c6380-etc-machine-id\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.436994 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-config-data\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.437071 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/920741ff-d4f2-495b-907c-d052b32c6380-etc-machine-id\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.437454 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-combined-ca-bundle\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.437498 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-db-sync-config-data\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.444566 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-combined-ca-bundle\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.444904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-config-data\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.448672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-scripts\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.452736 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-db-sync-config-data\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.453987 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbrwz\" (UniqueName: \"kubernetes.io/projected/920741ff-d4f2-495b-907c-d052b32c6380-kube-api-access-bbrwz\") pod \"cinder-db-sync-bszbk\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:39 crc kubenswrapper[4926]: I1007 21:13:39.568572 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bszbk" Oct 07 21:13:40 crc kubenswrapper[4926]: I1007 21:13:40.694616 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="920e2024-0025-4d13-a17c-a6993cb44eda" path="/var/lib/kubelet/pods/920e2024-0025-4d13-a17c-a6993cb44eda/volumes" Oct 07 21:13:41 crc kubenswrapper[4926]: I1007 21:13:41.690764 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.136:5353: connect: connection refused" Oct 07 21:13:44 crc kubenswrapper[4926]: I1007 21:13:44.408355 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.267146 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.267264 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.267459 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncdh67ch7dh98hf9h7h69h67hfdh546h58h5b9h6dh54fhc5hb7h684h5c6h688h5ch575h8fh658h66h655h598hchb7h64ch68h56fh5cfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6zjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-c59868cf7-9vlkn_openstack(08e6d63d-71ba-44d2-916f-ed3dd20ce845): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.269931 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-c59868cf7-9vlkn" podUID="08e6d63d-71ba-44d2-916f-ed3dd20ce845" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.279314 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.279371 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.279528 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n99h548h67chd7h59fh57h5dh68fh96hf7h694h5fchb6h5f8h594h67fh89h96h5c8h9ch5f7h696h5dh58chd8h59bh68hfbh66bh549h5h687q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dr2hc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5b9df8697c-5lvvt_openstack(ef793c0b-f866-4988-b559-e98432d41993): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.288258 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-5b9df8697c-5lvvt" podUID="ef793c0b-f866-4988-b559-e98432d41993" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.842329 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.842716 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 07 21:13:45 crc kubenswrapper[4926]: E1007 21:13:45.842844 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.12:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d5h7ch64h569h55fh5bbhbbh65ch9dh554h5ddh5c5h56chf9h74h87h78h66fh56ch59dh54h96h5d5h5c6h94hd7h669h574h574h648h5b5h59dq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7ql97,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f3476db5-0c85-4b2c-9378-ac237c3766fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:13:46 crc kubenswrapper[4926]: E1007 21:13:46.303013 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 07 21:13:46 crc kubenswrapper[4926]: E1007 21:13:46.303385 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 07 21:13:46 crc kubenswrapper[4926]: E1007 21:13:46.303902 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.12:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zpgjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-2sj9k_openstack(6bab0bbe-8fa5-451f-89fe-a22426ecdfe2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:13:46 crc kubenswrapper[4926]: E1007 21:13:46.305137 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-2sj9k" podUID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.434209 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"02582ac3-7293-4d98-9671-f1d68ad01013","Type":"ContainerDied","Data":"684ddc8d22b9838b53dcaaaf1128037e88dd5af5d9a9004c07f3237607f4fb53"} Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.434271 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="684ddc8d22b9838b53dcaaaf1128037e88dd5af5d9a9004c07f3237607f4fb53" Oct 07 21:13:46 crc kubenswrapper[4926]: E1007 21:13:46.444592 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-2sj9k" podUID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.594765 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.644849 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.665040 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.674393 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.703556 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-combined-ca-bundle\") pod \"02582ac3-7293-4d98-9671-f1d68ad01013\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.704155 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-config-data\") pod \"02582ac3-7293-4d98-9671-f1d68ad01013\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.704236 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02582ac3-7293-4d98-9671-f1d68ad01013-logs\") pod \"02582ac3-7293-4d98-9671-f1d68ad01013\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.704320 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-custom-prometheus-ca\") pod \"02582ac3-7293-4d98-9671-f1d68ad01013\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.704363 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfs9w\" (UniqueName: \"kubernetes.io/projected/02582ac3-7293-4d98-9671-f1d68ad01013-kube-api-access-qfs9w\") pod \"02582ac3-7293-4d98-9671-f1d68ad01013\" (UID: \"02582ac3-7293-4d98-9671-f1d68ad01013\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.706343 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02582ac3-7293-4d98-9671-f1d68ad01013-logs" (OuterVolumeSpecName: "logs") pod "02582ac3-7293-4d98-9671-f1d68ad01013" (UID: "02582ac3-7293-4d98-9671-f1d68ad01013"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.724154 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02582ac3-7293-4d98-9671-f1d68ad01013-kube-api-access-qfs9w" (OuterVolumeSpecName: "kube-api-access-qfs9w") pod "02582ac3-7293-4d98-9671-f1d68ad01013" (UID: "02582ac3-7293-4d98-9671-f1d68ad01013"). InnerVolumeSpecName "kube-api-access-qfs9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.737435 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02582ac3-7293-4d98-9671-f1d68ad01013" (UID: "02582ac3-7293-4d98-9671-f1d68ad01013"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.752303 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "02582ac3-7293-4d98-9671-f1d68ad01013" (UID: "02582ac3-7293-4d98-9671-f1d68ad01013"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.772278 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-config-data" (OuterVolumeSpecName: "config-data") pod "02582ac3-7293-4d98-9671-f1d68ad01013" (UID: "02582ac3-7293-4d98-9671-f1d68ad01013"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.805910 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr2hc\" (UniqueName: \"kubernetes.io/projected/ef793c0b-f866-4988-b559-e98432d41993-kube-api-access-dr2hc\") pod \"ef793c0b-f866-4988-b559-e98432d41993\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.805952 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08e6d63d-71ba-44d2-916f-ed3dd20ce845-logs\") pod \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806004 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-config\") pod \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806046 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6zjd\" (UniqueName: \"kubernetes.io/projected/08e6d63d-71ba-44d2-916f-ed3dd20ce845-kube-api-access-s6zjd\") pod \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806067 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-swift-storage-0\") pod \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806107 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-scripts\") pod \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806221 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef793c0b-f866-4988-b559-e98432d41993-horizon-secret-key\") pod \"ef793c0b-f866-4988-b559-e98432d41993\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806301 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-scripts\") pod \"ef793c0b-f866-4988-b559-e98432d41993\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806330 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-svc\") pod \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806344 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-nb\") pod \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806397 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-config-data\") pod \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806420 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08e6d63d-71ba-44d2-916f-ed3dd20ce845-horizon-secret-key\") pod \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\" (UID: \"08e6d63d-71ba-44d2-916f-ed3dd20ce845\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806445 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef793c0b-f866-4988-b559-e98432d41993-logs\") pod \"ef793c0b-f866-4988-b559-e98432d41993\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806489 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r86vm\" (UniqueName: \"kubernetes.io/projected/69fa95b0-9096-4383-bfbf-3af3c095c9ad-kube-api-access-r86vm\") pod \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806508 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-config-data\") pod \"ef793c0b-f866-4988-b559-e98432d41993\" (UID: \"ef793c0b-f866-4988-b559-e98432d41993\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806544 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-sb\") pod \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\" (UID: \"69fa95b0-9096-4383-bfbf-3af3c095c9ad\") " Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.806977 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02582ac3-7293-4d98-9671-f1d68ad01013-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.807002 4926 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.807014 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfs9w\" (UniqueName: \"kubernetes.io/projected/02582ac3-7293-4d98-9671-f1d68ad01013-kube-api-access-qfs9w\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.807028 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.807038 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02582ac3-7293-4d98-9671-f1d68ad01013-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.807759 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-scripts" (OuterVolumeSpecName: "scripts") pod "ef793c0b-f866-4988-b559-e98432d41993" (UID: "ef793c0b-f866-4988-b559-e98432d41993"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.808182 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08e6d63d-71ba-44d2-916f-ed3dd20ce845-logs" (OuterVolumeSpecName: "logs") pod "08e6d63d-71ba-44d2-916f-ed3dd20ce845" (UID: "08e6d63d-71ba-44d2-916f-ed3dd20ce845"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.808346 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef793c0b-f866-4988-b559-e98432d41993-logs" (OuterVolumeSpecName: "logs") pod "ef793c0b-f866-4988-b559-e98432d41993" (UID: "ef793c0b-f866-4988-b559-e98432d41993"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.809175 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-scripts" (OuterVolumeSpecName: "scripts") pod "08e6d63d-71ba-44d2-916f-ed3dd20ce845" (UID: "08e6d63d-71ba-44d2-916f-ed3dd20ce845"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.808970 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-config-data" (OuterVolumeSpecName: "config-data") pod "ef793c0b-f866-4988-b559-e98432d41993" (UID: "ef793c0b-f866-4988-b559-e98432d41993"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.809330 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-config-data" (OuterVolumeSpecName: "config-data") pod "08e6d63d-71ba-44d2-916f-ed3dd20ce845" (UID: "08e6d63d-71ba-44d2-916f-ed3dd20ce845"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.812078 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08e6d63d-71ba-44d2-916f-ed3dd20ce845-kube-api-access-s6zjd" (OuterVolumeSpecName: "kube-api-access-s6zjd") pod "08e6d63d-71ba-44d2-916f-ed3dd20ce845" (UID: "08e6d63d-71ba-44d2-916f-ed3dd20ce845"). InnerVolumeSpecName "kube-api-access-s6zjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.814386 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08e6d63d-71ba-44d2-916f-ed3dd20ce845-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "08e6d63d-71ba-44d2-916f-ed3dd20ce845" (UID: "08e6d63d-71ba-44d2-916f-ed3dd20ce845"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.816665 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef793c0b-f866-4988-b559-e98432d41993-kube-api-access-dr2hc" (OuterVolumeSpecName: "kube-api-access-dr2hc") pod "ef793c0b-f866-4988-b559-e98432d41993" (UID: "ef793c0b-f866-4988-b559-e98432d41993"). InnerVolumeSpecName "kube-api-access-dr2hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.818261 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69fa95b0-9096-4383-bfbf-3af3c095c9ad-kube-api-access-r86vm" (OuterVolumeSpecName: "kube-api-access-r86vm") pod "69fa95b0-9096-4383-bfbf-3af3c095c9ad" (UID: "69fa95b0-9096-4383-bfbf-3af3c095c9ad"). InnerVolumeSpecName "kube-api-access-r86vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.819046 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef793c0b-f866-4988-b559-e98432d41993-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ef793c0b-f866-4988-b559-e98432d41993" (UID: "ef793c0b-f866-4988-b559-e98432d41993"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.856885 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69fa95b0-9096-4383-bfbf-3af3c095c9ad" (UID: "69fa95b0-9096-4383-bfbf-3af3c095c9ad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.863863 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69fa95b0-9096-4383-bfbf-3af3c095c9ad" (UID: "69fa95b0-9096-4383-bfbf-3af3c095c9ad"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.865454 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-config" (OuterVolumeSpecName: "config") pod "69fa95b0-9096-4383-bfbf-3af3c095c9ad" (UID: "69fa95b0-9096-4383-bfbf-3af3c095c9ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.873775 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69fa95b0-9096-4383-bfbf-3af3c095c9ad" (UID: "69fa95b0-9096-4383-bfbf-3af3c095c9ad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.881350 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69fa95b0-9096-4383-bfbf-3af3c095c9ad" (UID: "69fa95b0-9096-4383-bfbf-3af3c095c9ad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.908936 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.908970 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.908984 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.908996 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909007 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08e6d63d-71ba-44d2-916f-ed3dd20ce845-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909018 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef793c0b-f866-4988-b559-e98432d41993-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909032 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r86vm\" (UniqueName: \"kubernetes.io/projected/69fa95b0-9096-4383-bfbf-3af3c095c9ad-kube-api-access-r86vm\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909043 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef793c0b-f866-4988-b559-e98432d41993-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909056 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909067 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr2hc\" (UniqueName: \"kubernetes.io/projected/ef793c0b-f866-4988-b559-e98432d41993-kube-api-access-dr2hc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909077 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08e6d63d-71ba-44d2-916f-ed3dd20ce845-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909088 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909102 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6zjd\" (UniqueName: \"kubernetes.io/projected/08e6d63d-71ba-44d2-916f-ed3dd20ce845-kube-api-access-s6zjd\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909116 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa95b0-9096-4383-bfbf-3af3c095c9ad-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909126 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e6d63d-71ba-44d2-916f-ed3dd20ce845-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:46 crc kubenswrapper[4926]: I1007 21:13:46.909137 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef793c0b-f866-4988-b559-e98432d41993-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.010429 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bf67dc754-22mcb"] Oct 07 21:13:47 crc kubenswrapper[4926]: W1007 21:13:47.016320 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6aa8d106_5d6e_4585_b368_e54382bed918.slice/crio-9239cae20f7fcbbb28459fcc4bc999b726077d4a72473b4b9ac8298b76bcb114 WatchSource:0}: Error finding container 9239cae20f7fcbbb28459fcc4bc999b726077d4a72473b4b9ac8298b76bcb114: Status 404 returned error can't find the container with id 9239cae20f7fcbbb28459fcc4bc999b726077d4a72473b4b9ac8298b76bcb114 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.017485 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-sjjzr"] Oct 07 21:13:47 crc kubenswrapper[4926]: W1007 21:13:47.018940 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3739cb7c_d23c_462e_a7ec_0bd07c2a0149.slice/crio-4c161a81298b890f572420a6a12514907b511623530dac51dacf294648350900 WatchSource:0}: Error finding container 4c161a81298b890f572420a6a12514907b511623530dac51dacf294648350900: Status 404 returned error can't find the container with id 4c161a81298b890f572420a6a12514907b511623530dac51dacf294648350900 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.190326 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bszbk"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.201924 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pvnmq"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.224046 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-858b4c475d-fcp7q"] Oct 07 21:13:47 crc kubenswrapper[4926]: W1007 21:13:47.381107 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod920741ff_d4f2_495b_907c_d052b32c6380.slice/crio-1a92f22d152f33f7ac2f3e52132f8b42f1eff528e04e5f1124fd264bf95f67db WatchSource:0}: Error finding container 1a92f22d152f33f7ac2f3e52132f8b42f1eff528e04e5f1124fd264bf95f67db: Status 404 returned error can't find the container with id 1a92f22d152f33f7ac2f3e52132f8b42f1eff528e04e5f1124fd264bf95f67db Oct 07 21:13:47 crc kubenswrapper[4926]: W1007 21:13:47.387765 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b74a118_4677_43dd_8506_dc66795fe8a1.slice/crio-0e0c83fdf771f9b59daacdb433c389a492a1ee6c751de8704c9f0d418539ca63 WatchSource:0}: Error finding container 0e0c83fdf771f9b59daacdb433c389a492a1ee6c751de8704c9f0d418539ca63: Status 404 returned error can't find the container with id 0e0c83fdf771f9b59daacdb433c389a492a1ee6c751de8704c9f0d418539ca63 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.468582 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerStarted","Data":"94424ffca67f31b70557f548664110fc9d36d2d1072473eeb4975b1fe157e99a"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.472218 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106","Type":"ContainerStarted","Data":"8ce288b5de3383c20e52fcd9d97a15938f2e6083f642bacf46c2d4bd8f14a52c"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.494076 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=4.7557661 podStartE2EDuration="24.494008492s" podCreationTimestamp="2025-10-07 21:13:23 +0000 UTC" firstStartedPulling="2025-10-07 21:13:26.546481778 +0000 UTC m=+1056.585062928" lastFinishedPulling="2025-10-07 21:13:46.28472416 +0000 UTC m=+1076.323305320" observedRunningTime="2025-10-07 21:13:47.491056956 +0000 UTC m=+1077.529638106" watchObservedRunningTime="2025-10-07 21:13:47.494008492 +0000 UTC m=+1077.532589642" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.496249 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dd68f6c57-k5dpz" event={"ID":"eff2499a-3993-4955-b32c-86883e6d649f","Type":"ContainerStarted","Data":"ab3ce47798ce625d5344024b99ace4d4077a7fb7405dbbfeaaf655f734c357b8"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.507360 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89","Type":"ContainerStarted","Data":"6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.511222 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-sjjzr" event={"ID":"6aa8d106-5d6e-4585-b368-e54382bed918","Type":"ContainerStarted","Data":"9239cae20f7fcbbb28459fcc4bc999b726077d4a72473b4b9ac8298b76bcb114"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.520020 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bszbk" event={"ID":"920741ff-d4f2-495b-907c-d052b32c6380","Type":"ContainerStarted","Data":"1a92f22d152f33f7ac2f3e52132f8b42f1eff528e04e5f1124fd264bf95f67db"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.530377 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-858b4c475d-fcp7q" event={"ID":"c42be0d4-2ced-4d32-84f4-49389e2a0e82","Type":"ContainerStarted","Data":"3e10458cd5eef2cda9f96a0b2ada94373c4d51518a88503e77d97228c9694ff8"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.532693 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=4.343420594 podStartE2EDuration="24.532673374s" podCreationTimestamp="2025-10-07 21:13:23 +0000 UTC" firstStartedPulling="2025-10-07 21:13:26.162366591 +0000 UTC m=+1056.200947741" lastFinishedPulling="2025-10-07 21:13:46.351619351 +0000 UTC m=+1076.390200521" observedRunningTime="2025-10-07 21:13:47.52496745 +0000 UTC m=+1077.563548600" watchObservedRunningTime="2025-10-07 21:13:47.532673374 +0000 UTC m=+1077.571254524" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.542743 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5f61f403-e0b6-45f1-a0a8-d073f1bdd190","Type":"ContainerStarted","Data":"8f832981f0da3122035fbafca4edf5f1b8f9fb1c8fa5295c9c67e5bc5c0135dc"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.543175 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-log" containerID="cri-o://bb71457dc900b962c61909e4cb2debcb69d33dd0779fb069ef52b617e460d485" gracePeriod=30 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.543302 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-httpd" containerID="cri-o://8f832981f0da3122035fbafca4edf5f1b8f9fb1c8fa5295c9c67e5bc5c0135dc" gracePeriod=30 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.575226 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9","Type":"ContainerStarted","Data":"d0e8d96abfea149ad86074b1041488140f4a593bc29102b16d12ddebf28fd52c"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.575401 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-log" containerID="cri-o://a13fa194e03bde11aa16a3ab57643b48bfbcb49b3c506611dc887dbab96cc31f" gracePeriod=30 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.575895 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-httpd" containerID="cri-o://d0e8d96abfea149ad86074b1041488140f4a593bc29102b16d12ddebf28fd52c" gracePeriod=30 Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.582517 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pvnmq" event={"ID":"0b74a118-4677-43dd-8506-dc66795fe8a1","Type":"ContainerStarted","Data":"0e0c83fdf771f9b59daacdb433c389a492a1ee6c751de8704c9f0d418539ca63"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.590158 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=23.590139961 podStartE2EDuration="23.590139961s" podCreationTimestamp="2025-10-07 21:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:47.571095529 +0000 UTC m=+1077.609676669" watchObservedRunningTime="2025-10-07 21:13:47.590139961 +0000 UTC m=+1077.628721111" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.590872 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bf67dc754-22mcb" event={"ID":"3739cb7c-d23c-462e-a7ec-0bd07c2a0149","Type":"ContainerStarted","Data":"4c161a81298b890f572420a6a12514907b511623530dac51dacf294648350900"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.596364 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b9df8697c-5lvvt" event={"ID":"ef793c0b-f866-4988-b559-e98432d41993","Type":"ContainerDied","Data":"952f6efcafc5500afb5315f269a2089b510f4199d66f4f35ad90c9522fc959d9"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.596403 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b9df8697c-5lvvt" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.605108 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c59868cf7-9vlkn" event={"ID":"08e6d63d-71ba-44d2-916f-ed3dd20ce845","Type":"ContainerDied","Data":"5abd251844d79c7c55563e6936e7a69cd989aad94b56f10237a213c937eb94e9"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.605219 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c59868cf7-9vlkn" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.606325 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=23.60629989 podStartE2EDuration="23.60629989s" podCreationTimestamp="2025-10-07 21:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:47.600904214 +0000 UTC m=+1077.639485364" watchObservedRunningTime="2025-10-07 21:13:47.60629989 +0000 UTC m=+1077.644881040" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.612794 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvgtl" event={"ID":"bf7199a1-5e63-4d53-93dd-5818fb4e0454","Type":"ContainerStarted","Data":"6484f4ad687b7564a9b1933e1bb2a4e5569ffe21f54c4e4f7254aa5bad110678"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.623093 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.623865 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.623961 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c6866c67-6cjbf" event={"ID":"69fa95b0-9096-4383-bfbf-3af3c095c9ad","Type":"ContainerDied","Data":"7aa1fb86119ed4a2dae48fefe86e6f9dc95d353056ece663eb238e7981e4bcce"} Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.624008 4926 scope.go:117] "RemoveContainer" containerID="3ea737c649028e281a9295bbcc0e840bb716c92299249b00ab2fdfcc847219e7" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.656144 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gvgtl" podStartSLOduration=3.495802905 podStartE2EDuration="23.656123086s" podCreationTimestamp="2025-10-07 21:13:24 +0000 UTC" firstStartedPulling="2025-10-07 21:13:26.192389622 +0000 UTC m=+1056.230970772" lastFinishedPulling="2025-10-07 21:13:46.352709803 +0000 UTC m=+1076.391290953" observedRunningTime="2025-10-07 21:13:47.631594314 +0000 UTC m=+1077.670175464" watchObservedRunningTime="2025-10-07 21:13:47.656123086 +0000 UTC m=+1077.694704236" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.701993 4926 scope.go:117] "RemoveContainer" containerID="4e63e2b1e30f68a064b41b00afdccc785c2b1218fbb90dbcd77b23dbb3a26e49" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.714985 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c59868cf7-9vlkn"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.740632 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c59868cf7-9vlkn"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.758982 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c6866c67-6cjbf"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.772327 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c6866c67-6cjbf"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.870351 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b9df8697c-5lvvt"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.887243 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b9df8697c-5lvvt"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.896646 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.906458 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.914776 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:47 crc kubenswrapper[4926]: E1007 21:13:47.915854 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="dnsmasq-dns" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.915879 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="dnsmasq-dns" Oct 07 21:13:47 crc kubenswrapper[4926]: E1007 21:13:47.915898 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api-log" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.915905 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api-log" Oct 07 21:13:47 crc kubenswrapper[4926]: E1007 21:13:47.915919 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="init" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.915926 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="init" Oct 07 21:13:47 crc kubenswrapper[4926]: E1007 21:13:47.915961 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.915969 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.916247 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.916267 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" containerName="dnsmasq-dns" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.916281 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api-log" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.917442 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.921221 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 07 21:13:47 crc kubenswrapper[4926]: I1007 21:13:47.923839 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.050544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-config-data\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.050952 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.050996 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.051028 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbs6k\" (UniqueName: \"kubernetes.io/projected/cda92361-c143-4887-8e3f-15509be2ea6c-kube-api-access-cbs6k\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.051170 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cda92361-c143-4887-8e3f-15509be2ea6c-logs\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.153398 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-config-data\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.153475 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.153531 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.153571 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbs6k\" (UniqueName: \"kubernetes.io/projected/cda92361-c143-4887-8e3f-15509be2ea6c-kube-api-access-cbs6k\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.153629 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cda92361-c143-4887-8e3f-15509be2ea6c-logs\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.154362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cda92361-c143-4887-8e3f-15509be2ea6c-logs\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.158590 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.160375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-config-data\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.161015 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.182848 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbs6k\" (UniqueName: \"kubernetes.io/projected/cda92361-c143-4887-8e3f-15509be2ea6c-kube-api-access-cbs6k\") pod \"watcher-api-0\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.243862 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.739912 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-dd68f6c57-k5dpz" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon-log" containerID="cri-o://ab3ce47798ce625d5344024b99ace4d4077a7fb7405dbbfeaaf655f734c357b8" gracePeriod=30 Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.748223 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-dd68f6c57-k5dpz" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon" containerID="cri-o://daaf00923859cc80f5a57d0906f93a0e82463f349d9d4e8220a975d53d69e99d" gracePeriod=30 Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.772954 4926 generic.go:334] "Generic (PLEG): container finished" podID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerID="8f832981f0da3122035fbafca4edf5f1b8f9fb1c8fa5295c9c67e5bc5c0135dc" exitCode=0 Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.773000 4926 generic.go:334] "Generic (PLEG): container finished" podID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerID="bb71457dc900b962c61909e4cb2debcb69d33dd0779fb069ef52b617e460d485" exitCode=143 Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.815818 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-pvnmq" podStartSLOduration=10.815795138 podStartE2EDuration="10.815795138s" podCreationTimestamp="2025-10-07 21:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:48.731959465 +0000 UTC m=+1078.770540635" watchObservedRunningTime="2025-10-07 21:13:48.815795138 +0000 UTC m=+1078.854376298" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.816466 4926 generic.go:334] "Generic (PLEG): container finished" podID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerID="d0e8d96abfea149ad86074b1041488140f4a593bc29102b16d12ddebf28fd52c" exitCode=0 Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.816576 4926 generic.go:334] "Generic (PLEG): container finished" podID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerID="a13fa194e03bde11aa16a3ab57643b48bfbcb49b3c506611dc887dbab96cc31f" exitCode=143 Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.833361 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" path="/var/lib/kubelet/pods/02582ac3-7293-4d98-9671-f1d68ad01013/volumes" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.841142 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08e6d63d-71ba-44d2-916f-ed3dd20ce845" path="/var/lib/kubelet/pods/08e6d63d-71ba-44d2-916f-ed3dd20ce845/volumes" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.841727 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69fa95b0-9096-4383-bfbf-3af3c095c9ad" path="/var/lib/kubelet/pods/69fa95b0-9096-4383-bfbf-3af3c095c9ad/volumes" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.842607 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef793c0b-f866-4988-b559-e98432d41993" path="/var/lib/kubelet/pods/ef793c0b-f866-4988-b559-e98432d41993/volumes" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843544 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pvnmq" event={"ID":"0b74a118-4677-43dd-8506-dc66795fe8a1","Type":"ContainerStarted","Data":"1872a1fee377837432fdcf378b981fb559f0eb2ac2fa9621d1cf9267d8ca2c8b"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843692 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bf67dc754-22mcb" event={"ID":"3739cb7c-d23c-462e-a7ec-0bd07c2a0149","Type":"ContainerStarted","Data":"d7ca7a40d73b191f4315170a6719b4141987aa15743f21b39f4748f7a868f914"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bf67dc754-22mcb" event={"ID":"3739cb7c-d23c-462e-a7ec-0bd07c2a0149","Type":"ContainerStarted","Data":"c998b9bb106c97402a01e919798bd57bd94c1b298710141166203f882e278cb6"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843812 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3476db5-0c85-4b2c-9378-ac237c3766fa","Type":"ContainerStarted","Data":"da72c77ea12e4db9b8a43736abc6763a069d1d927ecd4cebb44151e808ea21b2"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-858b4c475d-fcp7q" event={"ID":"c42be0d4-2ced-4d32-84f4-49389e2a0e82","Type":"ContainerStarted","Data":"24d20e9592465f61e79372e7504654b87c1c016070af6bf94931d1e9d69ea3a3"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843926 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-858b4c475d-fcp7q" event={"ID":"c42be0d4-2ced-4d32-84f4-49389e2a0e82","Type":"ContainerStarted","Data":"7312e6f4c78512ee6faf995134a1e2ac78bd79599b11e05a90d6acdba2efb427"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.843990 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dd68f6c57-k5dpz" event={"ID":"eff2499a-3993-4955-b32c-86883e6d649f","Type":"ContainerStarted","Data":"daaf00923859cc80f5a57d0906f93a0e82463f349d9d4e8220a975d53d69e99d"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.844049 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5f61f403-e0b6-45f1-a0a8-d073f1bdd190","Type":"ContainerDied","Data":"8f832981f0da3122035fbafca4edf5f1b8f9fb1c8fa5295c9c67e5bc5c0135dc"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.844107 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5f61f403-e0b6-45f1-a0a8-d073f1bdd190","Type":"ContainerDied","Data":"bb71457dc900b962c61909e4cb2debcb69d33dd0779fb069ef52b617e460d485"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.844181 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9","Type":"ContainerDied","Data":"d0e8d96abfea149ad86074b1041488140f4a593bc29102b16d12ddebf28fd52c"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.844258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9","Type":"ContainerDied","Data":"a13fa194e03bde11aa16a3ab57643b48bfbcb49b3c506611dc887dbab96cc31f"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.844365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-sjjzr" event={"ID":"6aa8d106-5d6e-4585-b368-e54382bed918","Type":"ContainerStarted","Data":"d0e4a79821442552cfbf800b1ce377ccb5a523fb80581655d9c7a01814118561"} Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.901704 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bf67dc754-22mcb" podStartSLOduration=16.90167153 podStartE2EDuration="16.90167153s" podCreationTimestamp="2025-10-07 21:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:48.865236753 +0000 UTC m=+1078.903817923" watchObservedRunningTime="2025-10-07 21:13:48.90167153 +0000 UTC m=+1078.940252680" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.930088 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-sjjzr" podStartSLOduration=14.930064404 podStartE2EDuration="14.930064404s" podCreationTimestamp="2025-10-07 21:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:48.886462379 +0000 UTC m=+1078.925043529" watchObservedRunningTime="2025-10-07 21:13:48.930064404 +0000 UTC m=+1078.968645554" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.937783 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-dd68f6c57-k5dpz" podStartSLOduration=6.438517535 podStartE2EDuration="21.937763588s" podCreationTimestamp="2025-10-07 21:13:27 +0000 UTC" firstStartedPulling="2025-10-07 21:13:31.068136409 +0000 UTC m=+1061.106717559" lastFinishedPulling="2025-10-07 21:13:46.567382452 +0000 UTC m=+1076.605963612" observedRunningTime="2025-10-07 21:13:48.90752852 +0000 UTC m=+1078.946109680" watchObservedRunningTime="2025-10-07 21:13:48.937763588 +0000 UTC m=+1078.976344738" Oct 07 21:13:48 crc kubenswrapper[4926]: I1007 21:13:48.950432 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.363492 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.383675 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.410347 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="02582ac3-7293-4d98-9671-f1d68ad01013" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.149:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.492737 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-config-data\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493011 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493052 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-logs\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493072 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s485t\" (UniqueName: \"kubernetes.io/projected/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-kube-api-access-s485t\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493096 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-logs\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493126 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-combined-ca-bundle\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493148 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-scripts\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493172 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-429wm\" (UniqueName: \"kubernetes.io/projected/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-kube-api-access-429wm\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493302 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-scripts\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493700 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-logs" (OuterVolumeSpecName: "logs") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493927 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-logs" (OuterVolumeSpecName: "logs") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493944 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-httpd-run\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.493975 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-internal-tls-certs\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494009 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-httpd-run\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494058 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-config-data\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494096 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-combined-ca-bundle\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494249 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494329 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494369 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-public-tls-certs\") pod \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\" (UID: \"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.494408 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\" (UID: \"5f61f403-e0b6-45f1-a0a8-d073f1bdd190\") " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.495034 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.495052 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.495061 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.495069 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.504376 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-kube-api-access-s485t" (OuterVolumeSpecName: "kube-api-access-s485t") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "kube-api-access-s485t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.506357 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.506355 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-scripts" (OuterVolumeSpecName: "scripts") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.506510 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-kube-api-access-429wm" (OuterVolumeSpecName: "kube-api-access-429wm") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "kube-api-access-429wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.508387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-scripts" (OuterVolumeSpecName: "scripts") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.514589 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.537077 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.549537 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.568948 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.575804 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-config-data" (OuterVolumeSpecName: "config-data") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.578309 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5f61f403-e0b6-45f1-a0a8-d073f1bdd190" (UID: "5f61f403-e0b6-45f1-a0a8-d073f1bdd190"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596856 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s485t\" (UniqueName: \"kubernetes.io/projected/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-kube-api-access-s485t\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596887 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596897 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596906 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-429wm\" (UniqueName: \"kubernetes.io/projected/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-kube-api-access-429wm\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596914 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596922 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596931 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596957 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596965 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f61f403-e0b6-45f1-a0a8-d073f1bdd190-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.596977 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.599237 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-config-data" (OuterVolumeSpecName: "config-data") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.617783 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" (UID: "4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.620338 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.626517 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.699210 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.699235 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.699248 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.699257 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.859124 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.862314 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5f61f403-e0b6-45f1-a0a8-d073f1bdd190","Type":"ContainerDied","Data":"e8585ba81c74f64cd5b2c8063e4a13c43b140638db05e2af25f4cdc350f21a3a"} Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.862372 4926 scope.go:117] "RemoveContainer" containerID="8f832981f0da3122035fbafca4edf5f1b8f9fb1c8fa5295c9c67e5bc5c0135dc" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.869019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerStarted","Data":"a5905b1793ca91f9fcf57228ef018d794db4a4aa6dd1c3db25c3c2e0a0fd468e"} Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.875303 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.875502 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9","Type":"ContainerDied","Data":"830d5a7dfaee87cc9978fe8830708b379543b7398ac99a1ddd4ff908c59a4265"} Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.881328 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cda92361-c143-4887-8e3f-15509be2ea6c","Type":"ContainerStarted","Data":"a1164cccbca41d3b737d3c8d69b0b509d4a7418f414341a91b87a5cf9ca80f86"} Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.881369 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cda92361-c143-4887-8e3f-15509be2ea6c","Type":"ContainerStarted","Data":"f202ff04cf4efd48faad902c5ff63b57a91f19b7e10029c50b0c8f8f6eedf869"} Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.923670 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-858b4c475d-fcp7q" podStartSLOduration=16.923649196 podStartE2EDuration="16.923649196s" podCreationTimestamp="2025-10-07 21:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:49.914001306 +0000 UTC m=+1079.952582456" watchObservedRunningTime="2025-10-07 21:13:49.923649196 +0000 UTC m=+1079.962230346" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.925435 4926 scope.go:117] "RemoveContainer" containerID="bb71457dc900b962c61909e4cb2debcb69d33dd0779fb069ef52b617e460d485" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.938560 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.959354 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.967577 4926 scope.go:117] "RemoveContainer" containerID="d0e8d96abfea149ad86074b1041488140f4a593bc29102b16d12ddebf28fd52c" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.971131 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.979311 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992111 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:49 crc kubenswrapper[4926]: E1007 21:13:49.992562 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-log" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992576 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-log" Oct 07 21:13:49 crc kubenswrapper[4926]: E1007 21:13:49.992609 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-httpd" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992615 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-httpd" Oct 07 21:13:49 crc kubenswrapper[4926]: E1007 21:13:49.992639 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-httpd" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992645 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-httpd" Oct 07 21:13:49 crc kubenswrapper[4926]: E1007 21:13:49.992658 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-log" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992664 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-log" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992959 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-log" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992979 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-httpd" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.992992 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" containerName="glance-log" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.993007 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" containerName="glance-httpd" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.994238 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:49 crc kubenswrapper[4926]: I1007 21:13:49.999280 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.020741 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.020933 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.021033 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lnm8t" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.021139 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.037409 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.038025 4926 scope.go:117] "RemoveContainer" containerID="a13fa194e03bde11aa16a3ab57643b48bfbcb49b3c506611dc887dbab96cc31f" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.039598 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.060374 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.065651 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.097830 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120690 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120752 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-config-data\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120778 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120809 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tblnt\" (UniqueName: \"kubernetes.io/projected/e47c8f35-5fce-4dc5-b86d-895606525d5f-kube-api-access-tblnt\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120843 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-logs\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120882 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120900 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120919 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120946 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-logs\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.120978 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.121000 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.121017 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhnpf\" (UniqueName: \"kubernetes.io/projected/1598a03a-83d6-45b1-830f-5ef8c37f177a-kube-api-access-xhnpf\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.121098 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.121623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.121844 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-scripts\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.121990 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.224670 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-config-data\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227587 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tblnt\" (UniqueName: \"kubernetes.io/projected/e47c8f35-5fce-4dc5-b86d-895606525d5f-kube-api-access-tblnt\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-logs\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227811 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227831 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227857 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227880 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-logs\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227927 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhnpf\" (UniqueName: \"kubernetes.io/projected/1598a03a-83d6-45b1-830f-5ef8c37f177a-kube-api-access-xhnpf\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.227961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.228023 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.228157 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.228224 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-scripts\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.228369 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.228434 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.228867 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.230123 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-logs\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.230773 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.231553 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.237673 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-logs\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.253547 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.254029 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhnpf\" (UniqueName: \"kubernetes.io/projected/1598a03a-83d6-45b1-830f-5ef8c37f177a-kube-api-access-xhnpf\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.255050 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.266447 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-config-data\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.274835 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-scripts\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.274903 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.275017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tblnt\" (UniqueName: \"kubernetes.io/projected/e47c8f35-5fce-4dc5-b86d-895606525d5f-kube-api-access-tblnt\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.275397 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.275758 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.275879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.278908 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.290344 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.348594 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.380923 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.388921 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.754164 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9" path="/var/lib/kubelet/pods/4704a63e-09e7-4ccc-80f1-7cbb61f3c0c9/volumes" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.778310 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f61f403-e0b6-45f1-a0a8-d073f1bdd190" path="/var/lib/kubelet/pods/5f61f403-e0b6-45f1-a0a8-d073f1bdd190/volumes" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.901655 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cda92361-c143-4887-8e3f-15509be2ea6c","Type":"ContainerStarted","Data":"1d027a665d9b467f3a603e5d2f1976b3df9eaf19a0ff1cee7c32edb0c1366aaf"} Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.903623 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.919702 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerStarted","Data":"e15617ef19af5594b54256816f4c297916f3694e1c8cf532720ac567c2db4d79"} Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.932863 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=3.932846071 podStartE2EDuration="3.932846071s" podCreationTimestamp="2025-10-07 21:13:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:50.928853065 +0000 UTC m=+1080.967434215" watchObservedRunningTime="2025-10-07 21:13:50.932846071 +0000 UTC m=+1080.971427221" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.941366 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.941445 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.973326 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=30.973302525 podStartE2EDuration="30.973302525s" podCreationTimestamp="2025-10-07 21:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:50.967455076 +0000 UTC m=+1081.006036226" watchObservedRunningTime="2025-10-07 21:13:50.973302525 +0000 UTC m=+1081.011883675" Oct 07 21:13:50 crc kubenswrapper[4926]: I1007 21:13:50.973424 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.198629 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:13:51 crc kubenswrapper[4926]: W1007 21:13:51.204152 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1598a03a_83d6_45b1_830f_5ef8c37f177a.slice/crio-0db1a2000501bde8fe124dff5e1449fb8311d8af48b5be77623e560f70e902f0 WatchSource:0}: Error finding container 0db1a2000501bde8fe124dff5e1449fb8311d8af48b5be77623e560f70e902f0: Status 404 returned error can't find the container with id 0db1a2000501bde8fe124dff5e1449fb8311d8af48b5be77623e560f70e902f0 Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.338845 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:13:51 crc kubenswrapper[4926]: W1007 21:13:51.348772 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47c8f35_5fce_4dc5_b86d_895606525d5f.slice/crio-37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69 WatchSource:0}: Error finding container 37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69: Status 404 returned error can't find the container with id 37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69 Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.969495 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1598a03a-83d6-45b1-830f-5ef8c37f177a","Type":"ContainerStarted","Data":"0db1a2000501bde8fe124dff5e1449fb8311d8af48b5be77623e560f70e902f0"} Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.978186 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e47c8f35-5fce-4dc5-b86d-895606525d5f","Type":"ContainerStarted","Data":"37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69"} Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.982000 4926 generic.go:334] "Generic (PLEG): container finished" podID="bf7199a1-5e63-4d53-93dd-5818fb4e0454" containerID="6484f4ad687b7564a9b1933e1bb2a4e5569ffe21f54c4e4f7254aa5bad110678" exitCode=0 Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.983981 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvgtl" event={"ID":"bf7199a1-5e63-4d53-93dd-5818fb4e0454","Type":"ContainerDied","Data":"6484f4ad687b7564a9b1933e1bb2a4e5569ffe21f54c4e4f7254aa5bad110678"} Oct 07 21:13:51 crc kubenswrapper[4926]: I1007 21:13:51.994089 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.005579 4926 generic.go:334] "Generic (PLEG): container finished" podID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerID="8ce288b5de3383c20e52fcd9d97a15938f2e6083f642bacf46c2d4bd8f14a52c" exitCode=1 Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.005656 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106","Type":"ContainerDied","Data":"8ce288b5de3383c20e52fcd9d97a15938f2e6083f642bacf46c2d4bd8f14a52c"} Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.006654 4926 scope.go:117] "RemoveContainer" containerID="8ce288b5de3383c20e52fcd9d97a15938f2e6083f642bacf46c2d4bd8f14a52c" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.040043 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e47c8f35-5fce-4dc5-b86d-895606525d5f","Type":"ContainerStarted","Data":"3f91018b9c9a7df718757a1a4f896e3345bb3fad2f99da4b8ad99a66b08ea24a"} Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.050243 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.050307 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1598a03a-83d6-45b1-830f-5ef8c37f177a","Type":"ContainerStarted","Data":"4fe99c09c2dea580c5940654bd065d473abfef0c0aba98c752107b97fd028994"} Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.243876 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.372546 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.372904 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.447372 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.447780 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.715446 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.843871 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-config-data\") pod \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.843962 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-combined-ca-bundle\") pod \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.844067 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-scripts\") pod \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.844125 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf7199a1-5e63-4d53-93dd-5818fb4e0454-logs\") pod \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.844250 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsld9\" (UniqueName: \"kubernetes.io/projected/bf7199a1-5e63-4d53-93dd-5818fb4e0454-kube-api-access-tsld9\") pod \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\" (UID: \"bf7199a1-5e63-4d53-93dd-5818fb4e0454\") " Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.851400 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf7199a1-5e63-4d53-93dd-5818fb4e0454-logs" (OuterVolumeSpecName: "logs") pod "bf7199a1-5e63-4d53-93dd-5818fb4e0454" (UID: "bf7199a1-5e63-4d53-93dd-5818fb4e0454"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.869431 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-scripts" (OuterVolumeSpecName: "scripts") pod "bf7199a1-5e63-4d53-93dd-5818fb4e0454" (UID: "bf7199a1-5e63-4d53-93dd-5818fb4e0454"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.874393 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf7199a1-5e63-4d53-93dd-5818fb4e0454-kube-api-access-tsld9" (OuterVolumeSpecName: "kube-api-access-tsld9") pod "bf7199a1-5e63-4d53-93dd-5818fb4e0454" (UID: "bf7199a1-5e63-4d53-93dd-5818fb4e0454"). InnerVolumeSpecName "kube-api-access-tsld9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.880347 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf7199a1-5e63-4d53-93dd-5818fb4e0454" (UID: "bf7199a1-5e63-4d53-93dd-5818fb4e0454"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.885688 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-config-data" (OuterVolumeSpecName: "config-data") pod "bf7199a1-5e63-4d53-93dd-5818fb4e0454" (UID: "bf7199a1-5e63-4d53-93dd-5818fb4e0454"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.946845 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.946878 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf7199a1-5e63-4d53-93dd-5818fb4e0454-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.946890 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsld9\" (UniqueName: \"kubernetes.io/projected/bf7199a1-5e63-4d53-93dd-5818fb4e0454-kube-api-access-tsld9\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.946915 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:53 crc kubenswrapper[4926]: I1007 21:13:53.946923 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7199a1-5e63-4d53-93dd-5818fb4e0454-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.061087 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvgtl" event={"ID":"bf7199a1-5e63-4d53-93dd-5818fb4e0454","Type":"ContainerDied","Data":"48bc7a3df5741c1c8edf8fa12deb8624247826e2324b62ff2b4a79fdd88950a9"} Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.062304 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48bc7a3df5741c1c8edf8fa12deb8624247826e2324b62ff2b4a79fdd88950a9" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.061340 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvgtl" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.063827 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1598a03a-83d6-45b1-830f-5ef8c37f177a","Type":"ContainerStarted","Data":"0e597c2d8da33aa0eab78483b2abf155682e10378c17a70a1f82b364957e5112"} Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.066029 4926 generic.go:334] "Generic (PLEG): container finished" podID="0b74a118-4677-43dd-8506-dc66795fe8a1" containerID="1872a1fee377837432fdcf378b981fb559f0eb2ac2fa9621d1cf9267d8ca2c8b" exitCode=0 Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.066098 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pvnmq" event={"ID":"0b74a118-4677-43dd-8506-dc66795fe8a1","Type":"ContainerDied","Data":"1872a1fee377837432fdcf378b981fb559f0eb2ac2fa9621d1cf9267d8ca2c8b"} Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.079289 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106","Type":"ContainerStarted","Data":"5a48bfe8a1c773f55bd00aafb384822293530918d1598830d6fc5fdaac57fc6a"} Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.096979 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.096956528 podStartE2EDuration="5.096956528s" podCreationTimestamp="2025-10-07 21:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:54.08528842 +0000 UTC m=+1084.123869570" watchObservedRunningTime="2025-10-07 21:13:54.096956528 +0000 UTC m=+1084.135537678" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.117840 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e47c8f35-5fce-4dc5-b86d-895606525d5f","Type":"ContainerStarted","Data":"0cc6049dec82fff93de6a1cff120ab74b4312141366363d17ead23276dfcc5dd"} Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.118415 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.156179 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-849bb7d7cd-xqmg2"] Oct 07 21:13:54 crc kubenswrapper[4926]: E1007 21:13:54.156935 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf7199a1-5e63-4d53-93dd-5818fb4e0454" containerName="placement-db-sync" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.156953 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf7199a1-5e63-4d53-93dd-5818fb4e0454" containerName="placement-db-sync" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.157131 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf7199a1-5e63-4d53-93dd-5818fb4e0454" containerName="placement-db-sync" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.158147 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.169798 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.170090 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.170217 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.170572 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.170677 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x7b2f" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.232338 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-849bb7d7cd-xqmg2"] Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.235332 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.2353023929999996 podStartE2EDuration="5.235302393s" podCreationTimestamp="2025-10-07 21:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:13:54.173701445 +0000 UTC m=+1084.212282595" watchObservedRunningTime="2025-10-07 21:13:54.235302393 +0000 UTC m=+1084.273883543" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.256478 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-public-tls-certs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.256803 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-logs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.257443 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-internal-tls-certs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.257944 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-scripts\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.258115 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-combined-ca-bundle\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.258311 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-config-data\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.258536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhhck\" (UniqueName: \"kubernetes.io/projected/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-kube-api-access-qhhck\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361106 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-public-tls-certs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361158 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-logs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-internal-tls-certs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361386 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-scripts\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361413 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-combined-ca-bundle\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361428 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-config-data\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.361452 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhhck\" (UniqueName: \"kubernetes.io/projected/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-kube-api-access-qhhck\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.362150 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-logs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.421432 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-config-data\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.422734 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhhck\" (UniqueName: \"kubernetes.io/projected/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-kube-api-access-qhhck\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.425642 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-scripts\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.436812 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-internal-tls-certs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.437179 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-combined-ca-bundle\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.439715 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af8a1d9-b36b-4f59-905c-12d2245ae3cf-public-tls-certs\") pod \"placement-849bb7d7cd-xqmg2\" (UID: \"1af8a1d9-b36b-4f59-905c-12d2245ae3cf\") " pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.498048 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.521241 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.569100 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.583836 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.583868 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.615426 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 07 21:13:54 crc kubenswrapper[4926]: I1007 21:13:54.619917 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:55 crc kubenswrapper[4926]: I1007 21:13:55.167383 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 07 21:13:55 crc kubenswrapper[4926]: I1007 21:13:55.188278 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:55 crc kubenswrapper[4926]: I1007 21:13:55.209258 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:13:55 crc kubenswrapper[4926]: I1007 21:13:55.239150 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.472515 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.520301 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-credential-keys\") pod \"0b74a118-4677-43dd-8506-dc66795fe8a1\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.520368 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-scripts\") pod \"0b74a118-4677-43dd-8506-dc66795fe8a1\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.520411 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-combined-ca-bundle\") pod \"0b74a118-4677-43dd-8506-dc66795fe8a1\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.521333 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-fernet-keys\") pod \"0b74a118-4677-43dd-8506-dc66795fe8a1\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.521552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4rfg\" (UniqueName: \"kubernetes.io/projected/0b74a118-4677-43dd-8506-dc66795fe8a1-kube-api-access-f4rfg\") pod \"0b74a118-4677-43dd-8506-dc66795fe8a1\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.521584 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-config-data\") pod \"0b74a118-4677-43dd-8506-dc66795fe8a1\" (UID: \"0b74a118-4677-43dd-8506-dc66795fe8a1\") " Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.528228 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0b74a118-4677-43dd-8506-dc66795fe8a1" (UID: "0b74a118-4677-43dd-8506-dc66795fe8a1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.528298 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b74a118-4677-43dd-8506-dc66795fe8a1-kube-api-access-f4rfg" (OuterVolumeSpecName: "kube-api-access-f4rfg") pod "0b74a118-4677-43dd-8506-dc66795fe8a1" (UID: "0b74a118-4677-43dd-8506-dc66795fe8a1"). InnerVolumeSpecName "kube-api-access-f4rfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.531422 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0b74a118-4677-43dd-8506-dc66795fe8a1" (UID: "0b74a118-4677-43dd-8506-dc66795fe8a1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.542890 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-scripts" (OuterVolumeSpecName: "scripts") pod "0b74a118-4677-43dd-8506-dc66795fe8a1" (UID: "0b74a118-4677-43dd-8506-dc66795fe8a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.552335 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b74a118-4677-43dd-8506-dc66795fe8a1" (UID: "0b74a118-4677-43dd-8506-dc66795fe8a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.555307 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-config-data" (OuterVolumeSpecName: "config-data") pod "0b74a118-4677-43dd-8506-dc66795fe8a1" (UID: "0b74a118-4677-43dd-8506-dc66795fe8a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.624369 4926 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.624406 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.624417 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.624428 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.624436 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4rfg\" (UniqueName: \"kubernetes.io/projected/0b74a118-4677-43dd-8506-dc66795fe8a1-kube-api-access-f4rfg\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:56 crc kubenswrapper[4926]: I1007 21:13:56.624446 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b74a118-4677-43dd-8506-dc66795fe8a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.160731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pvnmq" event={"ID":"0b74a118-4677-43dd-8506-dc66795fe8a1","Type":"ContainerDied","Data":"0e0c83fdf771f9b59daacdb433c389a492a1ee6c751de8704c9f0d418539ca63"} Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.160768 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pvnmq" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.160794 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e0c83fdf771f9b59daacdb433c389a492a1ee6c751de8704c9f0d418539ca63" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.164960 4926 generic.go:334] "Generic (PLEG): container finished" podID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerID="5a48bfe8a1c773f55bd00aafb384822293530918d1598830d6fc5fdaac57fc6a" exitCode=1 Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.165070 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106","Type":"ContainerDied","Data":"5a48bfe8a1c773f55bd00aafb384822293530918d1598830d6fc5fdaac57fc6a"} Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.165105 4926 scope.go:117] "RemoveContainer" containerID="8ce288b5de3383c20e52fcd9d97a15938f2e6083f642bacf46c2d4bd8f14a52c" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.165423 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" containerName="watcher-applier" containerID="cri-o://6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" gracePeriod=30 Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.622457 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-647f4cc4bb-zkt6b"] Oct 07 21:13:57 crc kubenswrapper[4926]: E1007 21:13:57.623261 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b74a118-4677-43dd-8506-dc66795fe8a1" containerName="keystone-bootstrap" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.623278 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b74a118-4677-43dd-8506-dc66795fe8a1" containerName="keystone-bootstrap" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.623491 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b74a118-4677-43dd-8506-dc66795fe8a1" containerName="keystone-bootstrap" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.624922 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.629965 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.630047 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.629969 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.631931 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-x9gv9" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.632106 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.632688 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.651937 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-647f4cc4bb-zkt6b"] Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.654721 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-config-data\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.654773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-fernet-keys\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.654862 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-credential-keys\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.654893 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-combined-ca-bundle\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.655049 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-internal-tls-certs\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.655134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-public-tls-certs\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.655368 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-scripts\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.655417 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdsl5\" (UniqueName: \"kubernetes.io/projected/434a0a19-516e-4685-be72-36976397fa42-kube-api-access-kdsl5\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.756881 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-config-data\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.756961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-fernet-keys\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.757079 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-credential-keys\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.757114 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-combined-ca-bundle\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.757148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-internal-tls-certs\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.757183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-public-tls-certs\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.757277 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-scripts\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.757315 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdsl5\" (UniqueName: \"kubernetes.io/projected/434a0a19-516e-4685-be72-36976397fa42-kube-api-access-kdsl5\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.764558 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-public-tls-certs\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.764951 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-credential-keys\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.765153 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-config-data\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.765230 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-fernet-keys\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.768834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-combined-ca-bundle\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.769186 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-internal-tls-certs\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.776673 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/434a0a19-516e-4685-be72-36976397fa42-scripts\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.784296 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdsl5\" (UniqueName: \"kubernetes.io/projected/434a0a19-516e-4685-be72-36976397fa42-kube-api-access-kdsl5\") pod \"keystone-647f4cc4bb-zkt6b\" (UID: \"434a0a19-516e-4685-be72-36976397fa42\") " pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:57 crc kubenswrapper[4926]: I1007 21:13:57.963163 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:13:58 crc kubenswrapper[4926]: I1007 21:13:58.107350 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:13:58 crc kubenswrapper[4926]: I1007 21:13:58.244907 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 07 21:13:58 crc kubenswrapper[4926]: I1007 21:13:58.251079 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.186583 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.326546 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393035 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-logs\") pod \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393070 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-combined-ca-bundle\") pod \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393174 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz569\" (UniqueName: \"kubernetes.io/projected/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-kube-api-access-pz569\") pod \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393274 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-custom-prometheus-ca\") pod \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393327 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-config-data\") pod \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\" (UID: \"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106\") " Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393527 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-logs" (OuterVolumeSpecName: "logs") pod "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" (UID: "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.393717 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.422882 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-kube-api-access-pz569" (OuterVolumeSpecName: "kube-api-access-pz569") pod "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" (UID: "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106"). InnerVolumeSpecName "kube-api-access-pz569". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.434344 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" (UID: "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.442106 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" (UID: "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.444114 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-config-data" (OuterVolumeSpecName: "config-data") pod "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" (UID: "fbac957e-ce41-4c6b-ada3-4ebfbeb0b106"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.501545 4926 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.501603 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.501617 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.501626 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz569\" (UniqueName: \"kubernetes.io/projected/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106-kube-api-access-pz569\") on node \"crc\" DevicePath \"\"" Oct 07 21:13:59 crc kubenswrapper[4926]: E1007 21:13:59.579825 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714 is running failed: container process not found" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Oct 07 21:13:59 crc kubenswrapper[4926]: E1007 21:13:59.580628 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714 is running failed: container process not found" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Oct 07 21:13:59 crc kubenswrapper[4926]: E1007 21:13:59.580889 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714 is running failed: container process not found" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Oct 07 21:13:59 crc kubenswrapper[4926]: E1007 21:13:59.580912 4926 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714 is running failed: container process not found" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" containerName="watcher-applier" Oct 07 21:13:59 crc kubenswrapper[4926]: I1007 21:13:59.975500 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-849bb7d7cd-xqmg2"] Oct 07 21:14:00 crc kubenswrapper[4926]: W1007 21:14:00.090911 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod434a0a19_516e_4685_be72_36976397fa42.slice/crio-15278e9f37a78c0c703d877e157706e1284571ec753dc7244b61ce918d547cf8 WatchSource:0}: Error finding container 15278e9f37a78c0c703d877e157706e1284571ec753dc7244b61ce918d547cf8: Status 404 returned error can't find the container with id 15278e9f37a78c0c703d877e157706e1284571ec753dc7244b61ce918d547cf8 Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.100013 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-647f4cc4bb-zkt6b"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.200791 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2sj9k" event={"ID":"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2","Type":"ContainerStarted","Data":"2ef232985aea731aaf6627828bc3c183b93481e16e8852f13acd59d680db339e"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.202574 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.203719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3476db5-0c85-4b2c-9378-ac237c3766fa","Type":"ContainerStarted","Data":"7b99091b1c0330be8de5a61020f454ab9528baff13f2e06388e32e18ab6dc813"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.205974 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-849bb7d7cd-xqmg2" event={"ID":"1af8a1d9-b36b-4f59-905c-12d2245ae3cf","Type":"ContainerStarted","Data":"5c1cf54ab1f42f86e2eb2e666b17fa3f2cc6a60bca29eecf4f1123f0cea008b2"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.210547 4926 generic.go:334] "Generic (PLEG): container finished" podID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" exitCode=0 Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.210661 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.210613 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89","Type":"ContainerDied","Data":"6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.210787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89","Type":"ContainerDied","Data":"1b656de4172b5067eabeb4e44b83d88d0d124ed413d4a930ee82196981cb9f13"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.210822 4926 scope.go:117] "RemoveContainer" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.212761 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-647f4cc4bb-zkt6b" event={"ID":"434a0a19-516e-4685-be72-36976397fa42","Type":"ContainerStarted","Data":"15278e9f37a78c0c703d877e157706e1284571ec753dc7244b61ce918d547cf8"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.214567 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.214551 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"fbac957e-ce41-4c6b-ada3-4ebfbeb0b106","Type":"ContainerDied","Data":"fd76c572cd22cfbdb51e46260814eba6f58b7aa9acd9e380cab294dfcc6947b1"} Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.223612 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2sj9k" podStartSLOduration=3.604049266 podStartE2EDuration="36.223588983s" podCreationTimestamp="2025-10-07 21:13:24 +0000 UTC" firstStartedPulling="2025-10-07 21:13:26.922569011 +0000 UTC m=+1056.961150161" lastFinishedPulling="2025-10-07 21:13:59.542108728 +0000 UTC m=+1089.580689878" observedRunningTime="2025-10-07 21:14:00.216721534 +0000 UTC m=+1090.255302684" watchObservedRunningTime="2025-10-07 21:14:00.223588983 +0000 UTC m=+1090.262170123" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.233418 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-595mb\" (UniqueName: \"kubernetes.io/projected/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-kube-api-access-595mb\") pod \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.233478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-logs\") pod \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.233535 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-combined-ca-bundle\") pod \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.233741 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-config-data\") pod \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\" (UID: \"0dea945e-c5cb-4d6f-b0df-fb2207cb0b89\") " Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.238347 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-logs" (OuterVolumeSpecName: "logs") pod "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" (UID: "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.243796 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-kube-api-access-595mb" (OuterVolumeSpecName: "kube-api-access-595mb") pod "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" (UID: "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89"). InnerVolumeSpecName "kube-api-access-595mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.341397 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-595mb\" (UniqueName: \"kubernetes.io/projected/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-kube-api-access-595mb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.341917 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.363316 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" (UID: "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.364333 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-config-data" (OuterVolumeSpecName: "config-data") pod "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" (UID: "0dea945e-c5cb-4d6f-b0df-fb2207cb0b89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.381658 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.381802 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.390845 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.390939 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.440583 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.444296 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.444322 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.450493 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.468979 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.522458 4926 scope.go:117] "RemoveContainer" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" Oct 07 21:14:00 crc kubenswrapper[4926]: E1007 21:14:00.522887 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714\": container with ID starting with 6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714 not found: ID does not exist" containerID="6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.522915 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714"} err="failed to get container status \"6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714\": rpc error: code = NotFound desc = could not find container \"6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714\": container with ID starting with 6b65240e57c7a78aa28a8fc5790c9cad87446b9dbfa47d08b38e6488a8493714 not found: ID does not exist" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.522937 4926 scope.go:117] "RemoveContainer" containerID="5a48bfe8a1c773f55bd00aafb384822293530918d1598830d6fc5fdaac57fc6a" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.537836 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.543839 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.566284 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.581042 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: E1007 21:14:00.581664 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerName="watcher-decision-engine" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.581687 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerName="watcher-decision-engine" Oct 07 21:14:00 crc kubenswrapper[4926]: E1007 21:14:00.581742 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" containerName="watcher-applier" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.581751 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" containerName="watcher-applier" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.581948 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerName="watcher-decision-engine" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.581965 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerName="watcher-decision-engine" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.581995 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" containerName="watcher-applier" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.582958 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.584930 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.611866 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.699654 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" path="/var/lib/kubelet/pods/fbac957e-ce41-4c6b-ada3-4ebfbeb0b106/volumes" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.700794 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.703401 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.713941 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: E1007 21:14:00.714480 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerName="watcher-decision-engine" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.714563 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbac957e-ce41-4c6b-ada3-4ebfbeb0b106" containerName="watcher-decision-engine" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.715620 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.720403 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.721179 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.748026 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.748075 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs4n8\" (UniqueName: \"kubernetes.io/projected/392cbe25-f961-43ce-bc58-453dba0de3a9-kube-api-access-fs4n8\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.748122 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/392cbe25-f961-43ce-bc58-453dba0de3a9-logs\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.748146 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.748207 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-config-data\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850145 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2feaba-0aff-4563-94f6-939db1594277-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850171 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs4n8\" (UniqueName: \"kubernetes.io/projected/392cbe25-f961-43ce-bc58-453dba0de3a9-kube-api-access-fs4n8\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850237 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/392cbe25-f961-43ce-bc58-453dba0de3a9-logs\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850270 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850324 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2feaba-0aff-4563-94f6-939db1594277-config-data\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850343 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-config-data\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850386 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2feaba-0aff-4563-94f6-939db1594277-logs\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850415 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz96k\" (UniqueName: \"kubernetes.io/projected/7d2feaba-0aff-4563-94f6-939db1594277-kube-api-access-lz96k\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.850942 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/392cbe25-f961-43ce-bc58-453dba0de3a9-logs\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.856680 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.856868 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-config-data\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.857135 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.868639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs4n8\" (UniqueName: \"kubernetes.io/projected/392cbe25-f961-43ce-bc58-453dba0de3a9-kube-api-access-fs4n8\") pod \"watcher-decision-engine-0\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.946011 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.952470 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2feaba-0aff-4563-94f6-939db1594277-config-data\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.952559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2feaba-0aff-4563-94f6-939db1594277-logs\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.952602 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz96k\" (UniqueName: \"kubernetes.io/projected/7d2feaba-0aff-4563-94f6-939db1594277-kube-api-access-lz96k\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.952673 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2feaba-0aff-4563-94f6-939db1594277-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.953769 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2feaba-0aff-4563-94f6-939db1594277-logs\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.956833 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2feaba-0aff-4563-94f6-939db1594277-config-data\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.959824 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2feaba-0aff-4563-94f6-939db1594277-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:00 crc kubenswrapper[4926]: I1007 21:14:00.979122 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz96k\" (UniqueName: \"kubernetes.io/projected/7d2feaba-0aff-4563-94f6-939db1594277-kube-api-access-lz96k\") pod \"watcher-applier-0\" (UID: \"7d2feaba-0aff-4563-94f6-939db1594277\") " pod="openstack/watcher-applier-0" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.036029 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.246959 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-849bb7d7cd-xqmg2" event={"ID":"1af8a1d9-b36b-4f59-905c-12d2245ae3cf","Type":"ContainerStarted","Data":"4fd3ed496059b19bf6a7d1147f43b291eb1f97d227855fa50f44b0fa4b1c28eb"} Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.252285 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-849bb7d7cd-xqmg2" event={"ID":"1af8a1d9-b36b-4f59-905c-12d2245ae3cf","Type":"ContainerStarted","Data":"b30fd409187d6fc19678828acd93fec45ad7800cd93b7e149eb769800d075fc0"} Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.252397 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.252438 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.276779 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-647f4cc4bb-zkt6b" event={"ID":"434a0a19-516e-4685-be72-36976397fa42","Type":"ContainerStarted","Data":"6151fe53b38b82663c7d8bc93bf5bd9d2e49c5f552531eb93389a5b96c616551"} Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.282876 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.286665 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.286716 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.286732 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.286750 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.297224 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-849bb7d7cd-xqmg2" podStartSLOduration=7.297179737 podStartE2EDuration="7.297179737s" podCreationTimestamp="2025-10-07 21:13:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:01.290466812 +0000 UTC m=+1091.329047972" watchObservedRunningTime="2025-10-07 21:14:01.297179737 +0000 UTC m=+1091.335760887" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.316149 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-647f4cc4bb-zkt6b" podStartSLOduration=4.316128517 podStartE2EDuration="4.316128517s" podCreationTimestamp="2025-10-07 21:13:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:01.312167532 +0000 UTC m=+1091.350748692" watchObservedRunningTime="2025-10-07 21:14:01.316128517 +0000 UTC m=+1091.354709667" Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.473766 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:14:01 crc kubenswrapper[4926]: I1007 21:14:01.643139 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.284303 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"7d2feaba-0aff-4563-94f6-939db1594277","Type":"ContainerStarted","Data":"a5c086d06f72f1302706384c2ef3e0b4806ffb51248ccb74dbe86f6b36d11758"} Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.284889 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"7d2feaba-0aff-4563-94f6-939db1594277","Type":"ContainerStarted","Data":"294d18183ac09e364a4af1e4301cbd2300fb48b770e79b2c63cb38d05e1d38cd"} Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.307332 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerStarted","Data":"9df6c237e40298415056a85462d5ab6cbab582e3b3ebd4b054f72d626353ad8c"} Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.307380 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerStarted","Data":"eee2d8930de42e94351b24102904da7250820c9feecac0c27babe13f169c075a"} Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.314979 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.314954051 podStartE2EDuration="2.314954051s" podCreationTimestamp="2025-10-07 21:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:02.301739918 +0000 UTC m=+1092.340321068" watchObservedRunningTime="2025-10-07 21:14:02.314954051 +0000 UTC m=+1092.353535201" Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.350851 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.350818352 podStartE2EDuration="2.350818352s" podCreationTimestamp="2025-10-07 21:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:02.334307513 +0000 UTC m=+1092.372888663" watchObservedRunningTime="2025-10-07 21:14:02.350818352 +0000 UTC m=+1092.389399502" Oct 07 21:14:02 crc kubenswrapper[4926]: I1007 21:14:02.697329 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dea945e-c5cb-4d6f-b0df-fb2207cb0b89" path="/var/lib/kubelet/pods/0dea945e-c5cb-4d6f-b0df-fb2207cb0b89/volumes" Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.210261 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.210314 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.299181 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.299467 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api-log" containerID="cri-o://a1164cccbca41d3b737d3c8d69b0b509d4a7418f414341a91b87a5cf9ca80f86" gracePeriod=30 Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.299918 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api" containerID="cri-o://1d027a665d9b467f3a603e5d2f1976b3df9eaf19a0ff1cee7c32edb0c1366aaf" gracePeriod=30 Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.332746 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.332777 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:14:03 crc kubenswrapper[4926]: I1007 21:14:03.451246 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-858b4c475d-fcp7q" podUID="c42be0d4-2ced-4d32-84f4-49389e2a0e82" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.162:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.162:8443: connect: connection refused" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.126300 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.293110 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.426611 4926 generic.go:334] "Generic (PLEG): container finished" podID="cda92361-c143-4887-8e3f-15509be2ea6c" containerID="1d027a665d9b467f3a603e5d2f1976b3df9eaf19a0ff1cee7c32edb0c1366aaf" exitCode=0 Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.426908 4926 generic.go:334] "Generic (PLEG): container finished" podID="cda92361-c143-4887-8e3f-15509be2ea6c" containerID="a1164cccbca41d3b737d3c8d69b0b509d4a7418f414341a91b87a5cf9ca80f86" exitCode=143 Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.427966 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cda92361-c143-4887-8e3f-15509be2ea6c","Type":"ContainerDied","Data":"1d027a665d9b467f3a603e5d2f1976b3df9eaf19a0ff1cee7c32edb0c1366aaf"} Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.427991 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cda92361-c143-4887-8e3f-15509be2ea6c","Type":"ContainerDied","Data":"a1164cccbca41d3b737d3c8d69b0b509d4a7418f414341a91b87a5cf9ca80f86"} Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.756914 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.816944 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-custom-prometheus-ca\") pod \"cda92361-c143-4887-8e3f-15509be2ea6c\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.817039 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-combined-ca-bundle\") pod \"cda92361-c143-4887-8e3f-15509be2ea6c\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.817155 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbs6k\" (UniqueName: \"kubernetes.io/projected/cda92361-c143-4887-8e3f-15509be2ea6c-kube-api-access-cbs6k\") pod \"cda92361-c143-4887-8e3f-15509be2ea6c\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.817175 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-config-data\") pod \"cda92361-c143-4887-8e3f-15509be2ea6c\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.817274 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cda92361-c143-4887-8e3f-15509be2ea6c-logs\") pod \"cda92361-c143-4887-8e3f-15509be2ea6c\" (UID: \"cda92361-c143-4887-8e3f-15509be2ea6c\") " Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.818084 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cda92361-c143-4887-8e3f-15509be2ea6c-logs" (OuterVolumeSpecName: "logs") pod "cda92361-c143-4887-8e3f-15509be2ea6c" (UID: "cda92361-c143-4887-8e3f-15509be2ea6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.824435 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cda92361-c143-4887-8e3f-15509be2ea6c-kube-api-access-cbs6k" (OuterVolumeSpecName: "kube-api-access-cbs6k") pod "cda92361-c143-4887-8e3f-15509be2ea6c" (UID: "cda92361-c143-4887-8e3f-15509be2ea6c"). InnerVolumeSpecName "kube-api-access-cbs6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.868377 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cda92361-c143-4887-8e3f-15509be2ea6c" (UID: "cda92361-c143-4887-8e3f-15509be2ea6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.925630 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.926086 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbs6k\" (UniqueName: \"kubernetes.io/projected/cda92361-c143-4887-8e3f-15509be2ea6c-kube-api-access-cbs6k\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.926103 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cda92361-c143-4887-8e3f-15509be2ea6c-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.925755 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.926262 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.969021 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-config-data" (OuterVolumeSpecName: "config-data") pod "cda92361-c143-4887-8e3f-15509be2ea6c" (UID: "cda92361-c143-4887-8e3f-15509be2ea6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:04 crc kubenswrapper[4926]: I1007 21:14:04.991959 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "cda92361-c143-4887-8e3f-15509be2ea6c" (UID: "cda92361-c143-4887-8e3f-15509be2ea6c"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.028474 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.028539 4926 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cda92361-c143-4887-8e3f-15509be2ea6c-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.317626 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.450648 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cda92361-c143-4887-8e3f-15509be2ea6c","Type":"ContainerDied","Data":"f202ff04cf4efd48faad902c5ff63b57a91f19b7e10029c50b0c8f8f6eedf869"} Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.450705 4926 scope.go:117] "RemoveContainer" containerID="1d027a665d9b467f3a603e5d2f1976b3df9eaf19a0ff1cee7c32edb0c1366aaf" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.450880 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.456510 4926 generic.go:334] "Generic (PLEG): container finished" podID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerID="9df6c237e40298415056a85462d5ab6cbab582e3b3ebd4b054f72d626353ad8c" exitCode=1 Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.456581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerDied","Data":"9df6c237e40298415056a85462d5ab6cbab582e3b3ebd4b054f72d626353ad8c"} Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.457698 4926 scope.go:117] "RemoveContainer" containerID="9df6c237e40298415056a85462d5ab6cbab582e3b3ebd4b054f72d626353ad8c" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.493346 4926 scope.go:117] "RemoveContainer" containerID="a1164cccbca41d3b737d3c8d69b0b509d4a7418f414341a91b87a5cf9ca80f86" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.525672 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.538260 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.546359 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:14:05 crc kubenswrapper[4926]: E1007 21:14:05.546834 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api-log" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.546849 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api-log" Oct 07 21:14:05 crc kubenswrapper[4926]: E1007 21:14:05.546877 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.546884 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.558899 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api-log" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.558945 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" containerName="watcher-api" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.564505 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.564650 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.567140 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.568267 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.568623 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.645702 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-config-data\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.645759 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-public-tls-certs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.645902 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.646103 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.646160 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.646247 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d69b703-1e74-4e92-943d-6c73f029020e-logs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.646292 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnslx\" (UniqueName: \"kubernetes.io/projected/3d69b703-1e74-4e92-943d-6c73f029020e-kube-api-access-tnslx\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.747942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.749664 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.749698 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.749762 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d69b703-1e74-4e92-943d-6c73f029020e-logs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.749788 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnslx\" (UniqueName: \"kubernetes.io/projected/3d69b703-1e74-4e92-943d-6c73f029020e-kube-api-access-tnslx\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.749888 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-config-data\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.749907 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-public-tls-certs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.750378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d69b703-1e74-4e92-943d-6c73f029020e-logs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.755669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-public-tls-certs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.755970 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.758520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.758586 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.763048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d69b703-1e74-4e92-943d-6c73f029020e-config-data\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.772620 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnslx\" (UniqueName: \"kubernetes.io/projected/3d69b703-1e74-4e92-943d-6c73f029020e-kube-api-access-tnslx\") pod \"watcher-api-0\" (UID: \"3d69b703-1e74-4e92-943d-6c73f029020e\") " pod="openstack/watcher-api-0" Oct 07 21:14:05 crc kubenswrapper[4926]: I1007 21:14:05.895642 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 07 21:14:06 crc kubenswrapper[4926]: I1007 21:14:06.013724 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:14:06 crc kubenswrapper[4926]: I1007 21:14:06.036983 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 07 21:14:06 crc kubenswrapper[4926]: I1007 21:14:06.471415 4926 generic.go:334] "Generic (PLEG): container finished" podID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" containerID="2ef232985aea731aaf6627828bc3c183b93481e16e8852f13acd59d680db339e" exitCode=0 Oct 07 21:14:06 crc kubenswrapper[4926]: I1007 21:14:06.471479 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2sj9k" event={"ID":"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2","Type":"ContainerDied","Data":"2ef232985aea731aaf6627828bc3c183b93481e16e8852f13acd59d680db339e"} Oct 07 21:14:06 crc kubenswrapper[4926]: I1007 21:14:06.484823 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerStarted","Data":"4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68"} Oct 07 21:14:06 crc kubenswrapper[4926]: I1007 21:14:06.700610 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cda92361-c143-4887-8e3f-15509be2ea6c" path="/var/lib/kubelet/pods/cda92361-c143-4887-8e3f-15509be2ea6c/volumes" Oct 07 21:14:07 crc kubenswrapper[4926]: I1007 21:14:07.602324 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:14:09 crc kubenswrapper[4926]: I1007 21:14:09.521068 4926 generic.go:334] "Generic (PLEG): container finished" podID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerID="4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68" exitCode=1 Oct 07 21:14:09 crc kubenswrapper[4926]: I1007 21:14:09.521149 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerDied","Data":"4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68"} Oct 07 21:14:09 crc kubenswrapper[4926]: I1007 21:14:09.521457 4926 scope.go:117] "RemoveContainer" containerID="9df6c237e40298415056a85462d5ab6cbab582e3b3ebd4b054f72d626353ad8c" Oct 07 21:14:09 crc kubenswrapper[4926]: I1007 21:14:09.522124 4926 scope.go:117] "RemoveContainer" containerID="4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68" Oct 07 21:14:09 crc kubenswrapper[4926]: E1007 21:14:09.522519 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:14:10 crc kubenswrapper[4926]: I1007 21:14:10.946593 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:10 crc kubenswrapper[4926]: I1007 21:14:10.946667 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:10 crc kubenswrapper[4926]: I1007 21:14:10.947590 4926 scope.go:117] "RemoveContainer" containerID="4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68" Oct 07 21:14:10 crc kubenswrapper[4926]: E1007 21:14:10.947879 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:14:11 crc kubenswrapper[4926]: I1007 21:14:11.036926 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 07 21:14:11 crc kubenswrapper[4926]: I1007 21:14:11.061048 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 07 21:14:11 crc kubenswrapper[4926]: I1007 21:14:11.587406 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 07 21:14:15 crc kubenswrapper[4926]: I1007 21:14:15.225974 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:14:15 crc kubenswrapper[4926]: E1007 21:14:15.719218 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 07 21:14:15 crc kubenswrapper[4926]: E1007 21:14:15.719278 4926 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.12:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 07 21:14:15 crc kubenswrapper[4926]: E1007 21:14:15.719417 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.12:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bbrwz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-bszbk_openstack(920741ff-d4f2-495b-907c-d052b32c6380): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 21:14:15 crc kubenswrapper[4926]: E1007 21:14:15.720859 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-bszbk" podUID="920741ff-d4f2-495b-907c-d052b32c6380" Oct 07 21:14:16 crc kubenswrapper[4926]: E1007 21:14:16.595704 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-bszbk" podUID="920741ff-d4f2-495b-907c-d052b32c6380" Oct 07 21:14:16 crc kubenswrapper[4926]: I1007 21:14:16.850931 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-858b4c475d-fcp7q" Oct 07 21:14:16 crc kubenswrapper[4926]: I1007 21:14:16.912233 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bf67dc754-22mcb"] Oct 07 21:14:16 crc kubenswrapper[4926]: I1007 21:14:16.912627 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bf67dc754-22mcb" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon-log" containerID="cri-o://c998b9bb106c97402a01e919798bd57bd94c1b298710141166203f882e278cb6" gracePeriod=30 Oct 07 21:14:16 crc kubenswrapper[4926]: I1007 21:14:16.912661 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bf67dc754-22mcb" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" containerID="cri-o://d7ca7a40d73b191f4315170a6719b4141987aa15743f21b39f4748f7a868f914" gracePeriod=30 Oct 07 21:14:17 crc kubenswrapper[4926]: I1007 21:14:17.605313 4926 generic.go:334] "Generic (PLEG): container finished" podID="6aa8d106-5d6e-4585-b368-e54382bed918" containerID="d0e4a79821442552cfbf800b1ce377ccb5a523fb80581655d9c7a01814118561" exitCode=0 Oct 07 21:14:17 crc kubenswrapper[4926]: I1007 21:14:17.605490 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-sjjzr" event={"ID":"6aa8d106-5d6e-4585-b368-e54382bed918","Type":"ContainerDied","Data":"d0e4a79821442552cfbf800b1ce377ccb5a523fb80581655d9c7a01814118561"} Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.518836 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.618283 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpgjw\" (UniqueName: \"kubernetes.io/projected/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-kube-api-access-zpgjw\") pod \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.618388 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-combined-ca-bundle\") pod \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.618562 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-db-sync-config-data\") pod \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\" (UID: \"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2\") " Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.623292 4926 generic.go:334] "Generic (PLEG): container finished" podID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerID="d7ca7a40d73b191f4315170a6719b4141987aa15743f21b39f4748f7a868f914" exitCode=0 Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.623352 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bf67dc754-22mcb" event={"ID":"3739cb7c-d23c-462e-a7ec-0bd07c2a0149","Type":"ContainerDied","Data":"d7ca7a40d73b191f4315170a6719b4141987aa15743f21b39f4748f7a868f914"} Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.626631 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" (UID: "6bab0bbe-8fa5-451f-89fe-a22426ecdfe2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.626724 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-kube-api-access-zpgjw" (OuterVolumeSpecName: "kube-api-access-zpgjw") pod "6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" (UID: "6bab0bbe-8fa5-451f-89fe-a22426ecdfe2"). InnerVolumeSpecName "kube-api-access-zpgjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.631864 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2sj9k" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.631819 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2sj9k" event={"ID":"6bab0bbe-8fa5-451f-89fe-a22426ecdfe2","Type":"ContainerDied","Data":"1e130eea95b2325a617ffa9fa5d752bdd74d577badba1dd35b22b07452e6dc29"} Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.632018 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e130eea95b2325a617ffa9fa5d752bdd74d577badba1dd35b22b07452e6dc29" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.667440 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" (UID: "6bab0bbe-8fa5-451f-89fe-a22426ecdfe2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.720618 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.720654 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:18 crc kubenswrapper[4926]: I1007 21:14:18.720669 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpgjw\" (UniqueName: \"kubernetes.io/projected/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2-kube-api-access-zpgjw\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.291473 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.331882 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnf5g\" (UniqueName: \"kubernetes.io/projected/6aa8d106-5d6e-4585-b368-e54382bed918-kube-api-access-hnf5g\") pod \"6aa8d106-5d6e-4585-b368-e54382bed918\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.332130 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-config\") pod \"6aa8d106-5d6e-4585-b368-e54382bed918\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.332165 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-combined-ca-bundle\") pod \"6aa8d106-5d6e-4585-b368-e54382bed918\" (UID: \"6aa8d106-5d6e-4585-b368-e54382bed918\") " Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.336860 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa8d106-5d6e-4585-b368-e54382bed918-kube-api-access-hnf5g" (OuterVolumeSpecName: "kube-api-access-hnf5g") pod "6aa8d106-5d6e-4585-b368-e54382bed918" (UID: "6aa8d106-5d6e-4585-b368-e54382bed918"). InnerVolumeSpecName "kube-api-access-hnf5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.356728 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-config" (OuterVolumeSpecName: "config") pod "6aa8d106-5d6e-4585-b368-e54382bed918" (UID: "6aa8d106-5d6e-4585-b368-e54382bed918"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.358438 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aa8d106-5d6e-4585-b368-e54382bed918" (UID: "6aa8d106-5d6e-4585-b368-e54382bed918"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.434902 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.435298 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa8d106-5d6e-4585-b368-e54382bed918-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.435310 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnf5g\" (UniqueName: \"kubernetes.io/projected/6aa8d106-5d6e-4585-b368-e54382bed918-kube-api-access-hnf5g\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.645081 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-sjjzr" event={"ID":"6aa8d106-5d6e-4585-b368-e54382bed918","Type":"ContainerDied","Data":"9239cae20f7fcbbb28459fcc4bc999b726077d4a72473b4b9ac8298b76bcb114"} Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.645131 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9239cae20f7fcbbb28459fcc4bc999b726077d4a72473b4b9ac8298b76bcb114" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.645222 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-sjjzr" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.660163 4926 generic.go:334] "Generic (PLEG): container finished" podID="eff2499a-3993-4955-b32c-86883e6d649f" containerID="daaf00923859cc80f5a57d0906f93a0e82463f349d9d4e8220a975d53d69e99d" exitCode=137 Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.660234 4926 generic.go:334] "Generic (PLEG): container finished" podID="eff2499a-3993-4955-b32c-86883e6d649f" containerID="ab3ce47798ce625d5344024b99ace4d4077a7fb7405dbbfeaaf655f734c357b8" exitCode=137 Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.660238 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dd68f6c57-k5dpz" event={"ID":"eff2499a-3993-4955-b32c-86883e6d649f","Type":"ContainerDied","Data":"daaf00923859cc80f5a57d0906f93a0e82463f349d9d4e8220a975d53d69e99d"} Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.660320 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dd68f6c57-k5dpz" event={"ID":"eff2499a-3993-4955-b32c-86883e6d649f","Type":"ContainerDied","Data":"ab3ce47798ce625d5344024b99ace4d4077a7fb7405dbbfeaaf655f734c357b8"} Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.876921 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-86dd45b86f-x4sv6"] Oct 07 21:14:19 crc kubenswrapper[4926]: E1007 21:14:19.877377 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" containerName="barbican-db-sync" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.877390 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" containerName="barbican-db-sync" Oct 07 21:14:19 crc kubenswrapper[4926]: E1007 21:14:19.877415 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa8d106-5d6e-4585-b368-e54382bed918" containerName="neutron-db-sync" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.877420 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa8d106-5d6e-4585-b368-e54382bed918" containerName="neutron-db-sync" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.877608 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa8d106-5d6e-4585-b368-e54382bed918" containerName="neutron-db-sync" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.877622 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" containerName="barbican-db-sync" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.878664 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.889708 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.889995 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-96wbv" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.890051 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.907494 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6774c5f56-pjjjg"] Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.909296 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.918300 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.929030 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-86dd45b86f-x4sv6"] Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.945769 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-config-data\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.945843 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afa6dc20-30d9-48f1-886f-6b436a2f0678-logs\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.945873 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qfn6\" (UniqueName: \"kubernetes.io/projected/5badd158-2c74-425b-bc80-95ebba967905-kube-api-access-5qfn6\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.945925 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-config-data-custom\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.945954 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhbhh\" (UniqueName: \"kubernetes.io/projected/afa6dc20-30d9-48f1-886f-6b436a2f0678-kube-api-access-hhbhh\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.946706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-config-data\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.946733 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-combined-ca-bundle\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.946761 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-combined-ca-bundle\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.946833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5badd158-2c74-425b-bc80-95ebba967905-logs\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.946856 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-config-data-custom\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:19 crc kubenswrapper[4926]: I1007 21:14:19.965362 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6774c5f56-pjjjg"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.004418 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6877db7f99-kdkmn"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.005999 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.044499 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6877db7f99-kdkmn"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047418 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-combined-ca-bundle\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047475 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f2bt\" (UniqueName: \"kubernetes.io/projected/1c93db67-c566-47b0-835c-53362f688e50-kube-api-access-6f2bt\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047518 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5badd158-2c74-425b-bc80-95ebba967905-logs\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-sb\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047556 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-config-data-custom\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047577 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-config-data\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047605 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afa6dc20-30d9-48f1-886f-6b436a2f0678-logs\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047621 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-config\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047646 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qfn6\" (UniqueName: \"kubernetes.io/projected/5badd158-2c74-425b-bc80-95ebba967905-kube-api-access-5qfn6\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-nb\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047695 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-svc\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047719 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-config-data-custom\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhbhh\" (UniqueName: \"kubernetes.io/projected/afa6dc20-30d9-48f1-886f-6b436a2f0678-kube-api-access-hhbhh\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-config-data\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-combined-ca-bundle\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.047827 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-swift-storage-0\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.048865 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5badd158-2c74-425b-bc80-95ebba967905-logs\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.051693 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afa6dc20-30d9-48f1-886f-6b436a2f0678-logs\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.055901 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-config-data-custom\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.057832 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-config-data-custom\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.059289 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-config-data\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.059816 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5badd158-2c74-425b-bc80-95ebba967905-combined-ca-bundle\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.060350 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-combined-ca-bundle\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.101588 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qfn6\" (UniqueName: \"kubernetes.io/projected/5badd158-2c74-425b-bc80-95ebba967905-kube-api-access-5qfn6\") pod \"barbican-keystone-listener-6774c5f56-pjjjg\" (UID: \"5badd158-2c74-425b-bc80-95ebba967905\") " pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.103396 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afa6dc20-30d9-48f1-886f-6b436a2f0678-config-data\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.103869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhbhh\" (UniqueName: \"kubernetes.io/projected/afa6dc20-30d9-48f1-886f-6b436a2f0678-kube-api-access-hhbhh\") pod \"barbican-worker-86dd45b86f-x4sv6\" (UID: \"afa6dc20-30d9-48f1-886f-6b436a2f0678\") " pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.149406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-config\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.149483 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-nb\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.149525 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-svc\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.149604 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-swift-storage-0\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.149658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f2bt\" (UniqueName: \"kubernetes.io/projected/1c93db67-c566-47b0-835c-53362f688e50-kube-api-access-6f2bt\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.149714 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-sb\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.158971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-config\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.161971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-swift-storage-0\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.162111 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-nb\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.166807 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-sb\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.168879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-svc\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.206994 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f2bt\" (UniqueName: \"kubernetes.io/projected/1c93db67-c566-47b0-835c-53362f688e50-kube-api-access-6f2bt\") pod \"dnsmasq-dns-6877db7f99-kdkmn\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.213257 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6877db7f99-kdkmn"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.213978 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.223665 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-86dd45b86f-x4sv6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.240641 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.269281 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-d546cd4db-lzl8j"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.270901 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.282980 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 07 21:14:20 crc kubenswrapper[4926]: E1007 21:14:20.319476 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 07 21:14:20 crc kubenswrapper[4926]: E1007 21:14:20.319637 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7ql97,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f3476db5-0c85-4b2c-9378-ac237c3766fa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 21:14:20 crc kubenswrapper[4926]: E1007 21:14:20.321315 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.347264 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d546cd4db-lzl8j"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.353433 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7b967f55-42zm8"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.355136 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data-custom\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.355185 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5d28ce0-4eeb-4408-9f66-d972bc438964-logs\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.355239 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.355270 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-combined-ca-bundle\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.355295 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6vt9\" (UniqueName: \"kubernetes.io/projected/b5d28ce0-4eeb-4408-9f66-d972bc438964-kube-api-access-t6vt9\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.355629 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.367262 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5df5c749bd-sp7b6"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.368870 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.377691 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4cxdd" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.377878 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.378056 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.379338 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7b967f55-42zm8"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.399540 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.405927 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df5c749bd-sp7b6"] Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456670 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-combined-ca-bundle\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456716 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-httpd-config\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456775 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456804 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data-custom\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456842 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5d28ce0-4eeb-4408-9f66-d972bc438964-logs\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456862 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-config\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456889 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-ovndb-tls-certs\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456910 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnthx\" (UniqueName: \"kubernetes.io/projected/e4537dd0-c22a-4879-9fa2-d02f3403172d-kube-api-access-lnthx\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456957 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-combined-ca-bundle\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456979 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6vt9\" (UniqueName: \"kubernetes.io/projected/b5d28ce0-4eeb-4408-9f66-d972bc438964-kube-api-access-t6vt9\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.456999 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-config\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.457038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hflj4\" (UniqueName: \"kubernetes.io/projected/334d53a8-d4f2-410f-8620-a489bb7d8261-kube-api-access-hflj4\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.457062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.457090 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-svc\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.459066 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5d28ce0-4eeb-4408-9f66-d972bc438964-logs\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.464974 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.470916 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-combined-ca-bundle\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.472957 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data-custom\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.520897 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6vt9\" (UniqueName: \"kubernetes.io/projected/b5d28ce0-4eeb-4408-9f66-d972bc438964-kube-api-access-t6vt9\") pod \"barbican-api-d546cd4db-lzl8j\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.558765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.558838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.558885 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-config\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.558931 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-ovndb-tls-certs\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.558964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnthx\" (UniqueName: \"kubernetes.io/projected/e4537dd0-c22a-4879-9fa2-d02f3403172d-kube-api-access-lnthx\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.559016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-config\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.559071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hflj4\" (UniqueName: \"kubernetes.io/projected/334d53a8-d4f2-410f-8620-a489bb7d8261-kube-api-access-hflj4\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.559101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.559130 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-svc\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.559157 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-combined-ca-bundle\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.559181 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-httpd-config\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.580516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.585056 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-httpd-config\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.586000 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-ovndb-tls-certs\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.587841 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-config\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.588082 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.592449 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-combined-ca-bundle\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.594161 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-config\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.594957 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.596149 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.599099 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hflj4\" (UniqueName: \"kubernetes.io/projected/334d53a8-d4f2-410f-8620-a489bb7d8261-kube-api-access-hflj4\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.609177 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnthx\" (UniqueName: \"kubernetes.io/projected/e4537dd0-c22a-4879-9fa2-d02f3403172d-kube-api-access-lnthx\") pod \"neutron-5df5c749bd-sp7b6\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.609967 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-svc\") pod \"dnsmasq-dns-5b7b967f55-42zm8\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.696461 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.720536 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="ceilometer-notification-agent" containerID="cri-o://da72c77ea12e4db9b8a43736abc6763a069d1d927ecd4cebb44151e808ea21b2" gracePeriod=30 Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.720685 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dd68f6c57-k5dpz" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.722414 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="sg-core" containerID="cri-o://7b99091b1c0330be8de5a61020f454ab9528baff13f2e06388e32e18ab6dc813" gracePeriod=30 Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.797327 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.830592 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dd68f6c57-k5dpz" event={"ID":"eff2499a-3993-4955-b32c-86883e6d649f","Type":"ContainerDied","Data":"1310059a53f84b291395ec24cd342a7492f32fbf2274671933fef56c30f36df3"} Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.830672 4926 scope.go:117] "RemoveContainer" containerID="daaf00923859cc80f5a57d0906f93a0e82463f349d9d4e8220a975d53d69e99d" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.834744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.869720 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eff2499a-3993-4955-b32c-86883e6d649f-logs\") pod \"eff2499a-3993-4955-b32c-86883e6d649f\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.872777 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eff2499a-3993-4955-b32c-86883e6d649f-logs" (OuterVolumeSpecName: "logs") pod "eff2499a-3993-4955-b32c-86883e6d649f" (UID: "eff2499a-3993-4955-b32c-86883e6d649f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.877256 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eff2499a-3993-4955-b32c-86883e6d649f-horizon-secret-key\") pod \"eff2499a-3993-4955-b32c-86883e6d649f\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.877329 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-scripts\") pod \"eff2499a-3993-4955-b32c-86883e6d649f\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.877495 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnmch\" (UniqueName: \"kubernetes.io/projected/eff2499a-3993-4955-b32c-86883e6d649f-kube-api-access-mnmch\") pod \"eff2499a-3993-4955-b32c-86883e6d649f\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.877581 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-config-data\") pod \"eff2499a-3993-4955-b32c-86883e6d649f\" (UID: \"eff2499a-3993-4955-b32c-86883e6d649f\") " Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.901006 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff2499a-3993-4955-b32c-86883e6d649f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "eff2499a-3993-4955-b32c-86883e6d649f" (UID: "eff2499a-3993-4955-b32c-86883e6d649f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.910876 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff2499a-3993-4955-b32c-86883e6d649f-kube-api-access-mnmch" (OuterVolumeSpecName: "kube-api-access-mnmch") pod "eff2499a-3993-4955-b32c-86883e6d649f" (UID: "eff2499a-3993-4955-b32c-86883e6d649f"). InnerVolumeSpecName "kube-api-access-mnmch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.939515 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-config-data" (OuterVolumeSpecName: "config-data") pod "eff2499a-3993-4955-b32c-86883e6d649f" (UID: "eff2499a-3993-4955-b32c-86883e6d649f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.944935 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-scripts" (OuterVolumeSpecName: "scripts") pod "eff2499a-3993-4955-b32c-86883e6d649f" (UID: "eff2499a-3993-4955-b32c-86883e6d649f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.982304 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnmch\" (UniqueName: \"kubernetes.io/projected/eff2499a-3993-4955-b32c-86883e6d649f-kube-api-access-mnmch\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.982347 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.982358 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eff2499a-3993-4955-b32c-86883e6d649f-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.982367 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eff2499a-3993-4955-b32c-86883e6d649f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:20 crc kubenswrapper[4926]: I1007 21:14:20.982376 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff2499a-3993-4955-b32c-86883e6d649f-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.078298 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-dd68f6c57-k5dpz"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.106825 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-dd68f6c57-k5dpz"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.142660 4926 scope.go:117] "RemoveContainer" containerID="ab3ce47798ce625d5344024b99ace4d4077a7fb7405dbbfeaaf655f734c357b8" Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.530177 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-86dd45b86f-x4sv6"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.554924 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6774c5f56-pjjjg"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.594536 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6877db7f99-kdkmn"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.617738 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.634131 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d546cd4db-lzl8j"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.735397 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-86dd45b86f-x4sv6" event={"ID":"afa6dc20-30d9-48f1-886f-6b436a2f0678","Type":"ContainerStarted","Data":"bd191d07b62c3904d5e51a3b3e5d0e3a6004fd64845869f6c093efb7a0d6dc11"} Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.738014 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" event={"ID":"5badd158-2c74-425b-bc80-95ebba967905","Type":"ContainerStarted","Data":"7c885516df7992db355f93a3b888f03e8e7545e885e401ac200cc7f98d800b40"} Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.739379 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d546cd4db-lzl8j" event={"ID":"b5d28ce0-4eeb-4408-9f66-d972bc438964","Type":"ContainerStarted","Data":"7dd70d15901324771575244211e64287b6e626872324527023a632759b2992c7"} Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.741027 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" event={"ID":"1c93db67-c566-47b0-835c-53362f688e50","Type":"ContainerStarted","Data":"189d32b9c2a4dcd132e08222917f71b1161f3383f344144f5e942225129bad2b"} Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.744700 4926 generic.go:334] "Generic (PLEG): container finished" podID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerID="7b99091b1c0330be8de5a61020f454ab9528baff13f2e06388e32e18ab6dc813" exitCode=2 Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.744779 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3476db5-0c85-4b2c-9378-ac237c3766fa","Type":"ContainerDied","Data":"7b99091b1c0330be8de5a61020f454ab9528baff13f2e06388e32e18ab6dc813"} Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.746559 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d69b703-1e74-4e92-943d-6c73f029020e","Type":"ContainerStarted","Data":"7c8ca2e740d06aea3c5a0ac317711ecc1ed404d9d6b2e98a96ac4edaa8319e59"} Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.845016 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df5c749bd-sp7b6"] Oct 07 21:14:21 crc kubenswrapper[4926]: I1007 21:14:21.882581 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7b967f55-42zm8"] Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.692534 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff2499a-3993-4955-b32c-86883e6d649f" path="/var/lib/kubelet/pods/eff2499a-3993-4955-b32c-86883e6d649f/volumes" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.706038 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b64c68dfc-k895v"] Oct 07 21:14:22 crc kubenswrapper[4926]: E1007 21:14:22.706550 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.706572 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon" Oct 07 21:14:22 crc kubenswrapper[4926]: E1007 21:14:22.706606 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon-log" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.706615 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon-log" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.706836 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon-log" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.706873 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff2499a-3993-4955-b32c-86883e6d649f" containerName="horizon" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.708162 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.717831 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.723806 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.736883 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b64c68dfc-k895v"] Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.750353 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-config\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.750443 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bdn7\" (UniqueName: \"kubernetes.io/projected/c8be70c1-8342-414d-ba4b-098d40277bf6-kube-api-access-4bdn7\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.750530 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-internal-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.750617 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-ovndb-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.750679 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-public-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.750753 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-httpd-config\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.751148 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-combined-ca-bundle\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.756758 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df5c749bd-sp7b6" event={"ID":"e4537dd0-c22a-4879-9fa2-d02f3403172d","Type":"ContainerStarted","Data":"383b5077ce69a5fe963cf6a75b6ac4a4dcf5d2afa9e458f126cd5305e1699950"} Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.757844 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" event={"ID":"334d53a8-d4f2-410f-8620-a489bb7d8261","Type":"ContainerStarted","Data":"365f0c6abc27888799fc58a3ff9cf1a39c7d61c1d4441728af389fb6b870556f"} Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.853539 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-config\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.853905 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bdn7\" (UniqueName: \"kubernetes.io/projected/c8be70c1-8342-414d-ba4b-098d40277bf6-kube-api-access-4bdn7\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.853940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-internal-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.853969 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-ovndb-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.853997 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-public-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.854045 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-httpd-config\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.854140 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-combined-ca-bundle\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.861911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-httpd-config\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.861989 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-internal-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.862165 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-public-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.862225 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-config\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.862669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-combined-ca-bundle\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.868845 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8be70c1-8342-414d-ba4b-098d40277bf6-ovndb-tls-certs\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:22 crc kubenswrapper[4926]: I1007 21:14:22.886008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bdn7\" (UniqueName: \"kubernetes.io/projected/c8be70c1-8342-414d-ba4b-098d40277bf6-kube-api-access-4bdn7\") pod \"neutron-b64c68dfc-k895v\" (UID: \"c8be70c1-8342-414d-ba4b-098d40277bf6\") " pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:23 crc kubenswrapper[4926]: I1007 21:14:23.024945 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:23 crc kubenswrapper[4926]: I1007 21:14:23.366961 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bf67dc754-22mcb" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.161:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.161:8443: connect: connection refused" Oct 07 21:14:23 crc kubenswrapper[4926]: I1007 21:14:23.612663 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b64c68dfc-k895v"] Oct 07 21:14:23 crc kubenswrapper[4926]: I1007 21:14:23.780892 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b64c68dfc-k895v" event={"ID":"c8be70c1-8342-414d-ba4b-098d40277bf6","Type":"ContainerStarted","Data":"be06ed5221a66a7fc2026a230582d27911a0ae485a24163c0ccac003065949ac"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.815371 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df5c749bd-sp7b6" event={"ID":"e4537dd0-c22a-4879-9fa2-d02f3403172d","Type":"ContainerStarted","Data":"eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.815637 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df5c749bd-sp7b6" event={"ID":"e4537dd0-c22a-4879-9fa2-d02f3403172d","Type":"ContainerStarted","Data":"f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.816770 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.821616 4926 generic.go:334] "Generic (PLEG): container finished" podID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerID="5b31e20cf3b6d4e484a1e2595386f0f383cb4de11a5c468db2158bd10cb58258" exitCode=0 Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.822477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" event={"ID":"334d53a8-d4f2-410f-8620-a489bb7d8261","Type":"ContainerDied","Data":"5b31e20cf3b6d4e484a1e2595386f0f383cb4de11a5c468db2158bd10cb58258"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.863451 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d546cd4db-lzl8j" event={"ID":"b5d28ce0-4eeb-4408-9f66-d972bc438964","Type":"ContainerStarted","Data":"87633ac3281fa3d94463173bc2df32fe2e7eec6f4eed36deb19d3f0426f24876"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.863760 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d546cd4db-lzl8j" event={"ID":"b5d28ce0-4eeb-4408-9f66-d972bc438964","Type":"ContainerStarted","Data":"100e249606f04d81891685b0722b1d158b92756d193466fcbb473d99a9a79041"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.864679 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.864777 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.874374 4926 generic.go:334] "Generic (PLEG): container finished" podID="1c93db67-c566-47b0-835c-53362f688e50" containerID="8113bb0785dbb80579d9084e9e190804398a62a9788e78d11c64e9a80fe811cc" exitCode=0 Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.874687 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" event={"ID":"1c93db67-c566-47b0-835c-53362f688e50","Type":"ContainerDied","Data":"8113bb0785dbb80579d9084e9e190804398a62a9788e78d11c64e9a80fe811cc"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.908985 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5df5c749bd-sp7b6" podStartSLOduration=4.908969462 podStartE2EDuration="4.908969462s" podCreationTimestamp="2025-10-07 21:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:24.855531451 +0000 UTC m=+1114.894112601" watchObservedRunningTime="2025-10-07 21:14:24.908969462 +0000 UTC m=+1114.947550612" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.916066 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b64c68dfc-k895v" event={"ID":"c8be70c1-8342-414d-ba4b-098d40277bf6","Type":"ContainerStarted","Data":"481f13b81492dffe75ad0f5b020fafe0d2640d610c8befa668cc72780a047c17"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.933421 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.933446 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b64c68dfc-k895v" event={"ID":"c8be70c1-8342-414d-ba4b-098d40277bf6","Type":"ContainerStarted","Data":"4e4b5ffd473d1dffc6af941b4516d3992cb67b2b26d793158d84e649d7a1f813"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.943696 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-d546cd4db-lzl8j" podStartSLOduration=4.943674993 podStartE2EDuration="4.943674993s" podCreationTimestamp="2025-10-07 21:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:24.909091415 +0000 UTC m=+1114.947672565" watchObservedRunningTime="2025-10-07 21:14:24.943674993 +0000 UTC m=+1114.982256143" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.948482 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d69b703-1e74-4e92-943d-6c73f029020e","Type":"ContainerStarted","Data":"06a89bc998ba503dd8e16ae2cace7d69b0408086133260bdfe4101642b57292a"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.948672 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d69b703-1e74-4e92-943d-6c73f029020e","Type":"ContainerStarted","Data":"576a1a48da22ec6aee541f5b9f8f759825813325deb3801cec867cd879066e6b"} Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.950874 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 07 21:14:24 crc kubenswrapper[4926]: I1007 21:14:24.979025 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b64c68dfc-k895v" podStartSLOduration=2.979001502 podStartE2EDuration="2.979001502s" podCreationTimestamp="2025-10-07 21:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:24.977852209 +0000 UTC m=+1115.016433369" watchObservedRunningTime="2025-10-07 21:14:24.979001502 +0000 UTC m=+1115.017582662" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.053810 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=20.053786329 podStartE2EDuration="20.053786329s" podCreationTimestamp="2025-10-07 21:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:25.017840132 +0000 UTC m=+1115.056421282" watchObservedRunningTime="2025-10-07 21:14:25.053786329 +0000 UTC m=+1115.092367479" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.496038 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.526603 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-config\") pod \"1c93db67-c566-47b0-835c-53362f688e50\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.526680 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-nb\") pod \"1c93db67-c566-47b0-835c-53362f688e50\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.526752 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-sb\") pod \"1c93db67-c566-47b0-835c-53362f688e50\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.527345 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-svc\") pod \"1c93db67-c566-47b0-835c-53362f688e50\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.527446 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f2bt\" (UniqueName: \"kubernetes.io/projected/1c93db67-c566-47b0-835c-53362f688e50-kube-api-access-6f2bt\") pod \"1c93db67-c566-47b0-835c-53362f688e50\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.527468 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-swift-storage-0\") pod \"1c93db67-c566-47b0-835c-53362f688e50\" (UID: \"1c93db67-c566-47b0-835c-53362f688e50\") " Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.547450 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c93db67-c566-47b0-835c-53362f688e50-kube-api-access-6f2bt" (OuterVolumeSpecName: "kube-api-access-6f2bt") pod "1c93db67-c566-47b0-835c-53362f688e50" (UID: "1c93db67-c566-47b0-835c-53362f688e50"). InnerVolumeSpecName "kube-api-access-6f2bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.552842 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c93db67-c566-47b0-835c-53362f688e50" (UID: "1c93db67-c566-47b0-835c-53362f688e50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.566290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-config" (OuterVolumeSpecName: "config") pod "1c93db67-c566-47b0-835c-53362f688e50" (UID: "1c93db67-c566-47b0-835c-53362f688e50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.566680 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c93db67-c566-47b0-835c-53362f688e50" (UID: "1c93db67-c566-47b0-835c-53362f688e50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.631433 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1c93db67-c566-47b0-835c-53362f688e50" (UID: "1c93db67-c566-47b0-835c-53362f688e50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.643237 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f2bt\" (UniqueName: \"kubernetes.io/projected/1c93db67-c566-47b0-835c-53362f688e50-kube-api-access-6f2bt\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.643282 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.643295 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.643308 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.643319 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.682169 4926 scope.go:117] "RemoveContainer" containerID="4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.713656 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c93db67-c566-47b0-835c-53362f688e50" (UID: "1c93db67-c566-47b0-835c-53362f688e50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.747893 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c93db67-c566-47b0-835c-53362f688e50-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.899473 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.900124 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.996635 4926 generic.go:334] "Generic (PLEG): container finished" podID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerID="da72c77ea12e4db9b8a43736abc6763a069d1d927ecd4cebb44151e808ea21b2" exitCode=0 Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.996696 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3476db5-0c85-4b2c-9378-ac237c3766fa","Type":"ContainerDied","Data":"da72c77ea12e4db9b8a43736abc6763a069d1d927ecd4cebb44151e808ea21b2"} Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.996725 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3476db5-0c85-4b2c-9378-ac237c3766fa","Type":"ContainerDied","Data":"d010e8d17d60781e94e6e378d9cfe69ec2dc204a92406284bc6afbe9755f4298"} Oct 07 21:14:25 crc kubenswrapper[4926]: I1007 21:14:25.996736 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d010e8d17d60781e94e6e378d9cfe69ec2dc204a92406284bc6afbe9755f4298" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.002437 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.004489 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" event={"ID":"334d53a8-d4f2-410f-8620-a489bb7d8261","Type":"ContainerStarted","Data":"debb1ed18035e660e47af53e537283270b1b3486e56b09532a96eca19b69100a"} Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.005455 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.018460 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.018848 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6877db7f99-kdkmn" event={"ID":"1c93db67-c566-47b0-835c-53362f688e50","Type":"ContainerDied","Data":"189d32b9c2a4dcd132e08222917f71b1161f3383f344144f5e942225129bad2b"} Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.024310 4926 scope.go:117] "RemoveContainer" containerID="8113bb0785dbb80579d9084e9e190804398a62a9788e78d11c64e9a80fe811cc" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.053859 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-scripts\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.053934 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-run-httpd\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.053962 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-log-httpd\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.053982 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-sg-core-conf-yaml\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.054105 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-config-data\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.054171 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ql97\" (UniqueName: \"kubernetes.io/projected/f3476db5-0c85-4b2c-9378-ac237c3766fa-kube-api-access-7ql97\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.054269 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-combined-ca-bundle\") pod \"f3476db5-0c85-4b2c-9378-ac237c3766fa\" (UID: \"f3476db5-0c85-4b2c-9378-ac237c3766fa\") " Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.060944 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.061146 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.084988 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" podStartSLOduration=6.084969431 podStartE2EDuration="6.084969431s" podCreationTimestamp="2025-10-07 21:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:26.074591241 +0000 UTC m=+1116.113172391" watchObservedRunningTime="2025-10-07 21:14:26.084969431 +0000 UTC m=+1116.123550581" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.100688 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-scripts" (OuterVolumeSpecName: "scripts") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.101422 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3476db5-0c85-4b2c-9378-ac237c3766fa-kube-api-access-7ql97" (OuterVolumeSpecName: "kube-api-access-7ql97") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "kube-api-access-7ql97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.132338 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.151160 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-config-data" (OuterVolumeSpecName: "config-data") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.158021 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.158137 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.158226 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3476db5-0c85-4b2c-9378-ac237c3766fa-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.158297 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.158358 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.158488 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ql97\" (UniqueName: \"kubernetes.io/projected/f3476db5-0c85-4b2c-9378-ac237c3766fa-kube-api-access-7ql97\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.191315 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3476db5-0c85-4b2c-9378-ac237c3766fa" (UID: "f3476db5-0c85-4b2c-9378-ac237c3766fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.229929 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6877db7f99-kdkmn"] Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.250872 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6877db7f99-kdkmn"] Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.266804 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3476db5-0c85-4b2c-9378-ac237c3766fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.723497 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c93db67-c566-47b0-835c-53362f688e50" path="/var/lib/kubelet/pods/1c93db67-c566-47b0-835c-53362f688e50/volumes" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.868636 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6696d67754-xn5b4"] Oct 07 21:14:26 crc kubenswrapper[4926]: E1007 21:14:26.869093 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c93db67-c566-47b0-835c-53362f688e50" containerName="init" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.869111 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c93db67-c566-47b0-835c-53362f688e50" containerName="init" Oct 07 21:14:26 crc kubenswrapper[4926]: E1007 21:14:26.869136 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="sg-core" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.869143 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="sg-core" Oct 07 21:14:26 crc kubenswrapper[4926]: E1007 21:14:26.869207 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="ceilometer-notification-agent" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.869214 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="ceilometer-notification-agent" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.869397 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c93db67-c566-47b0-835c-53362f688e50" containerName="init" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.869431 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="sg-core" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.869452 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" containerName="ceilometer-notification-agent" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.870577 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.872486 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.873127 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.879261 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6696d67754-xn5b4"] Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.907387 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/watcher-api-0" podUID="3d69b703-1e74-4e92-943d-6c73f029020e" containerName="watcher-api-log" probeResult="failure" output="Get \"https://10.217.0.173:9322/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985372 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-config-data-custom\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985446 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-logs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985546 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-config-data\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk42z\" (UniqueName: \"kubernetes.io/projected/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-kube-api-access-sk42z\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985632 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-combined-ca-bundle\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985697 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-internal-tls-certs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:26 crc kubenswrapper[4926]: I1007 21:14:26.985731 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-public-tls-certs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.026412 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.026594 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090200 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk42z\" (UniqueName: \"kubernetes.io/projected/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-kube-api-access-sk42z\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090299 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-combined-ca-bundle\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090408 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-internal-tls-certs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090461 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-public-tls-certs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-config-data-custom\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090798 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-logs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.090825 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-config-data\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.094332 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.095524 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-logs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.109654 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.115225 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-internal-tls-certs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.115442 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-config-data-custom\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.115925 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.117990 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-public-tls-certs\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.118594 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-combined-ca-bundle\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.121368 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.125423 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-config-data\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.128405 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.128743 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.134947 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk42z\" (UniqueName: \"kubernetes.io/projected/d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c-kube-api-access-sk42z\") pod \"barbican-api-6696d67754-xn5b4\" (UID: \"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c\") " pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.135319 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.196255 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.197814 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.197907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-log-httpd\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.197938 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.197958 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-run-httpd\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.197999 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvp5k\" (UniqueName: \"kubernetes.io/projected/c6ca2cc7-f855-4bbf-aab6-657da7506355-kube-api-access-bvp5k\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.198016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-scripts\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.198123 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-config-data\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-log-httpd\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300738 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-run-httpd\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300786 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvp5k\" (UniqueName: \"kubernetes.io/projected/c6ca2cc7-f855-4bbf-aab6-657da7506355-kube-api-access-bvp5k\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300805 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-scripts\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-config-data\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.300905 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.301654 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-run-httpd\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.302294 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-log-httpd\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.307952 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-config-data\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.309352 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-scripts\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.309520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.309785 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.321746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvp5k\" (UniqueName: \"kubernetes.io/projected/c6ca2cc7-f855-4bbf-aab6-657da7506355-kube-api-access-bvp5k\") pod \"ceilometer-0\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.522816 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.557687 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:14:27 crc kubenswrapper[4926]: I1007 21:14:27.570398 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-849bb7d7cd-xqmg2" Oct 07 21:14:28 crc kubenswrapper[4926]: I1007 21:14:28.039636 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerStarted","Data":"08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9"} Oct 07 21:14:28 crc kubenswrapper[4926]: I1007 21:14:28.174830 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 07 21:14:28 crc kubenswrapper[4926]: I1007 21:14:28.703487 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3476db5-0c85-4b2c-9378-ac237c3766fa" path="/var/lib/kubelet/pods/f3476db5-0c85-4b2c-9378-ac237c3766fa/volumes" Oct 07 21:14:28 crc kubenswrapper[4926]: I1007 21:14:28.758532 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6696d67754-xn5b4"] Oct 07 21:14:28 crc kubenswrapper[4926]: I1007 21:14:28.954928 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:29 crc kubenswrapper[4926]: I1007 21:14:29.079921 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerStarted","Data":"e3fbccb17bd4c2092b32d5255dd04a4496efd72ba175fc4b6ef5689353b5b269"} Oct 07 21:14:29 crc kubenswrapper[4926]: I1007 21:14:29.081447 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6696d67754-xn5b4" event={"ID":"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c","Type":"ContainerStarted","Data":"bf4b0c0c0a027c6aac4ce64e3716cd4913edcf82160de8c29581e236b1f1398f"} Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.096965 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" event={"ID":"5badd158-2c74-425b-bc80-95ebba967905","Type":"ContainerStarted","Data":"eb9139dac887dbe21c38f7faf4f6977c7ef4a3e5ac71a3090561308fa97efd45"} Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.098856 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6696d67754-xn5b4" event={"ID":"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c","Type":"ContainerStarted","Data":"bd0bbd5b0e043083ae610786e844f26319e765705a0d3834dd5e1438b00cd9b0"} Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.100106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-86dd45b86f-x4sv6" event={"ID":"afa6dc20-30d9-48f1-886f-6b436a2f0678","Type":"ContainerStarted","Data":"23a6590847945d3d57ffa460ff7951c02dc55bd2c72195a7a2e2b339296938be"} Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.837439 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.903665 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-544499d945-t6gsp"] Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.903893 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-544499d945-t6gsp" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerName="dnsmasq-dns" containerID="cri-o://f09c937d588e21c7a130574801d8780b483960c6786825e920895c89cd6b25c0" gracePeriod=10 Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.946545 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:30 crc kubenswrapper[4926]: I1007 21:14:30.947710 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.042542 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" probeResult="failure" output="" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.134408 4926 generic.go:334] "Generic (PLEG): container finished" podID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerID="08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9" exitCode=1 Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.134471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerDied","Data":"08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9"} Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.134511 4926 scope.go:117] "RemoveContainer" containerID="4c24f5f5530b9cc0b9cc736fa75704640efebac52deedb01a16f153b956dac68" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.136277 4926 scope.go:117] "RemoveContainer" containerID="08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9" Oct 07 21:14:31 crc kubenswrapper[4926]: E1007 21:14:31.136571 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.152486 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6696d67754-xn5b4" event={"ID":"d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c","Type":"ContainerStarted","Data":"1699ea7420267050caf1a10f36c19412dad6dffdd4abfebad4839ca9c6d2446b"} Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.152590 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.152614 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.194967 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-86dd45b86f-x4sv6" event={"ID":"afa6dc20-30d9-48f1-886f-6b436a2f0678","Type":"ContainerStarted","Data":"4fbbcb73a1fcb51d875986f3016ccacca364754d4a8df6ad1564a84138406e69"} Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.205632 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6696d67754-xn5b4" podStartSLOduration=5.205607176 podStartE2EDuration="5.205607176s" podCreationTimestamp="2025-10-07 21:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:31.182793788 +0000 UTC m=+1121.221374938" watchObservedRunningTime="2025-10-07 21:14:31.205607176 +0000 UTC m=+1121.244188316" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.206272 4926 generic.go:334] "Generic (PLEG): container finished" podID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerID="f09c937d588e21c7a130574801d8780b483960c6786825e920895c89cd6b25c0" exitCode=0 Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.206347 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544499d945-t6gsp" event={"ID":"efa18a6d-e823-4b88-9e0c-30e0361f6b98","Type":"ContainerDied","Data":"f09c937d588e21c7a130574801d8780b483960c6786825e920895c89cd6b25c0"} Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.208352 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" event={"ID":"5badd158-2c74-425b-bc80-95ebba967905","Type":"ContainerStarted","Data":"06e5720dfd1498aa293e02486f11246387b364690385edabcde07645a255fbaa"} Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.238662 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-86dd45b86f-x4sv6" podStartSLOduration=5.533986735 podStartE2EDuration="12.238635669s" podCreationTimestamp="2025-10-07 21:14:19 +0000 UTC" firstStartedPulling="2025-10-07 21:14:21.545712794 +0000 UTC m=+1111.584293944" lastFinishedPulling="2025-10-07 21:14:28.250361728 +0000 UTC m=+1118.288942878" observedRunningTime="2025-10-07 21:14:31.208848389 +0000 UTC m=+1121.247429539" watchObservedRunningTime="2025-10-07 21:14:31.238635669 +0000 UTC m=+1121.277216819" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.316507 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6774c5f56-pjjjg" podStartSLOduration=5.627464331 podStartE2EDuration="12.316482904s" podCreationTimestamp="2025-10-07 21:14:19 +0000 UTC" firstStartedPulling="2025-10-07 21:14:21.562104477 +0000 UTC m=+1111.600685627" lastFinishedPulling="2025-10-07 21:14:28.25112305 +0000 UTC m=+1118.289704200" observedRunningTime="2025-10-07 21:14:31.231790251 +0000 UTC m=+1121.270371421" watchObservedRunningTime="2025-10-07 21:14:31.316482904 +0000 UTC m=+1121.355064054" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.506145 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-647f4cc4bb-zkt6b" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.721341 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.756901 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-svc\") pod \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.756972 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-sb\") pod \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.757074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r9hm\" (UniqueName: \"kubernetes.io/projected/efa18a6d-e823-4b88-9e0c-30e0361f6b98-kube-api-access-8r9hm\") pod \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.757137 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-swift-storage-0\") pod \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.757169 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-config\") pod \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.757207 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-nb\") pod \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\" (UID: \"efa18a6d-e823-4b88-9e0c-30e0361f6b98\") " Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.823153 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efa18a6d-e823-4b88-9e0c-30e0361f6b98-kube-api-access-8r9hm" (OuterVolumeSpecName: "kube-api-access-8r9hm") pod "efa18a6d-e823-4b88-9e0c-30e0361f6b98" (UID: "efa18a6d-e823-4b88-9e0c-30e0361f6b98"). InnerVolumeSpecName "kube-api-access-8r9hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.859722 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r9hm\" (UniqueName: \"kubernetes.io/projected/efa18a6d-e823-4b88-9e0c-30e0361f6b98-kube-api-access-8r9hm\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.936723 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "efa18a6d-e823-4b88-9e0c-30e0361f6b98" (UID: "efa18a6d-e823-4b88-9e0c-30e0361f6b98"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.937776 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "efa18a6d-e823-4b88-9e0c-30e0361f6b98" (UID: "efa18a6d-e823-4b88-9e0c-30e0361f6b98"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.938237 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "efa18a6d-e823-4b88-9e0c-30e0361f6b98" (UID: "efa18a6d-e823-4b88-9e0c-30e0361f6b98"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.966396 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.968720 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.968818 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.968493 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "efa18a6d-e823-4b88-9e0c-30e0361f6b98" (UID: "efa18a6d-e823-4b88-9e0c-30e0361f6b98"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:31 crc kubenswrapper[4926]: I1007 21:14:31.977391 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-config" (OuterVolumeSpecName: "config") pod "efa18a6d-e823-4b88-9e0c-30e0361f6b98" (UID: "efa18a6d-e823-4b88-9e0c-30e0361f6b98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.072352 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.072390 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efa18a6d-e823-4b88-9e0c-30e0361f6b98-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.218125 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544499d945-t6gsp" event={"ID":"efa18a6d-e823-4b88-9e0c-30e0361f6b98","Type":"ContainerDied","Data":"db74c6e0c10427cdbd6cf117d43966a96bbed0fcb713a078492a2b14d1a0c9ba"} Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.218207 4926 scope.go:117] "RemoveContainer" containerID="f09c937d588e21c7a130574801d8780b483960c6786825e920895c89cd6b25c0" Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.218303 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544499d945-t6gsp" Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.236016 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerStarted","Data":"e13486028f9002ce9e8c63687bf53bc78ba59386ef5f69d83f0214b2f5cc1c0d"} Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.236083 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerStarted","Data":"13f3f5c396375822cd81862cf7afb57fe3c0480d24f4b1198b1df0097f952269"} Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.259944 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-544499d945-t6gsp"] Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.270568 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-544499d945-t6gsp"] Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.285621 4926 scope.go:117] "RemoveContainer" containerID="4701694df8bf30c6a8c6493b44b997b38a6d956cf0b6c5f94d3cdc9f6e116d9e" Oct 07 21:14:32 crc kubenswrapper[4926]: I1007 21:14:32.691737 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" path="/var/lib/kubelet/pods/efa18a6d-e823-4b88-9e0c-30e0361f6b98/volumes" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.008457 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.209391 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.209646 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.253478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerStarted","Data":"a457ee3746b3889fd816cdb00c10ab622eeed901dc9247a2f9efcc9cf9e6f2fb"} Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.255492 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bszbk" event={"ID":"920741ff-d4f2-495b-907c-d052b32c6380","Type":"ContainerStarted","Data":"dc64d9036eb9b61d494dbf7d6dee1f340a89e9c9f72316ae6cdbbce7aa276c07"} Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.273499 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-bszbk" podStartSLOduration=10.808751067 podStartE2EDuration="54.273482739s" podCreationTimestamp="2025-10-07 21:13:39 +0000 UTC" firstStartedPulling="2025-10-07 21:13:47.393125974 +0000 UTC m=+1077.431707134" lastFinishedPulling="2025-10-07 21:14:30.857857656 +0000 UTC m=+1120.896438806" observedRunningTime="2025-10-07 21:14:33.269921687 +0000 UTC m=+1123.308502837" watchObservedRunningTime="2025-10-07 21:14:33.273482739 +0000 UTC m=+1123.312063889" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.367060 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bf67dc754-22mcb" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.161:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.161:8443: connect: connection refused" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.518919 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.723182 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 07 21:14:33 crc kubenswrapper[4926]: E1007 21:14:33.723597 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerName="dnsmasq-dns" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.723612 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerName="dnsmasq-dns" Oct 07 21:14:33 crc kubenswrapper[4926]: E1007 21:14:33.723650 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerName="init" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.723657 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerName="init" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.727529 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa18a6d-e823-4b88-9e0c-30e0361f6b98" containerName="dnsmasq-dns" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.728176 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.734457 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.734687 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.734796 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-dz5nb" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.748796 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.806420 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-openstack-config-secret\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.806489 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-openstack-config\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.806647 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzxfk\" (UniqueName: \"kubernetes.io/projected/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-kube-api-access-nzxfk\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.806704 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.909004 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzxfk\" (UniqueName: \"kubernetes.io/projected/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-kube-api-access-nzxfk\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.909084 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.909156 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-openstack-config-secret\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.909187 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-openstack-config\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.910061 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-openstack-config\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.916119 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.916410 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-openstack-config-secret\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:33 crc kubenswrapper[4926]: I1007 21:14:33.927856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzxfk\" (UniqueName: \"kubernetes.io/projected/2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15-kube-api-access-nzxfk\") pod \"openstackclient\" (UID: \"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15\") " pod="openstack/openstackclient" Oct 07 21:14:34 crc kubenswrapper[4926]: I1007 21:14:34.050824 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 21:14:34 crc kubenswrapper[4926]: I1007 21:14:34.591213 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 21:14:35 crc kubenswrapper[4926]: I1007 21:14:35.290345 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15","Type":"ContainerStarted","Data":"53c52cbae4eab4d32f4e0bb31a778b293bfaccd3b01dc26c5174490662396d0f"} Oct 07 21:14:35 crc kubenswrapper[4926]: I1007 21:14:35.292945 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerStarted","Data":"eadcf578f81834730354002bb692272ab0d732f8f30a1a8fc687543ca9404af9"} Oct 07 21:14:35 crc kubenswrapper[4926]: I1007 21:14:35.293308 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:14:35 crc kubenswrapper[4926]: I1007 21:14:35.908820 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 07 21:14:35 crc kubenswrapper[4926]: I1007 21:14:35.915368 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 07 21:14:35 crc kubenswrapper[4926]: I1007 21:14:35.935530 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.906289112 podStartE2EDuration="8.93551144s" podCreationTimestamp="2025-10-07 21:14:27 +0000 UTC" firstStartedPulling="2025-10-07 21:14:28.975035599 +0000 UTC m=+1119.013616749" lastFinishedPulling="2025-10-07 21:14:34.004257927 +0000 UTC m=+1124.042839077" observedRunningTime="2025-10-07 21:14:35.319641658 +0000 UTC m=+1125.358222818" watchObservedRunningTime="2025-10-07 21:14:35.93551144 +0000 UTC m=+1125.974092590" Oct 07 21:14:39 crc kubenswrapper[4926]: I1007 21:14:39.076312 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:39 crc kubenswrapper[4926]: I1007 21:14:39.328763 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6696d67754-xn5b4" Oct 07 21:14:39 crc kubenswrapper[4926]: I1007 21:14:39.384802 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-d546cd4db-lzl8j"] Oct 07 21:14:39 crc kubenswrapper[4926]: I1007 21:14:39.385023 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-d546cd4db-lzl8j" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api-log" containerID="cri-o://100e249606f04d81891685b0722b1d158b92756d193466fcbb473d99a9a79041" gracePeriod=30 Oct 07 21:14:39 crc kubenswrapper[4926]: I1007 21:14:39.385585 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-d546cd4db-lzl8j" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api" containerID="cri-o://87633ac3281fa3d94463173bc2df32fe2e7eec6f4eed36deb19d3f0426f24876" gracePeriod=30 Oct 07 21:14:40 crc kubenswrapper[4926]: I1007 21:14:40.350216 4926 generic.go:334] "Generic (PLEG): container finished" podID="920741ff-d4f2-495b-907c-d052b32c6380" containerID="dc64d9036eb9b61d494dbf7d6dee1f340a89e9c9f72316ae6cdbbce7aa276c07" exitCode=0 Oct 07 21:14:40 crc kubenswrapper[4926]: I1007 21:14:40.350381 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bszbk" event={"ID":"920741ff-d4f2-495b-907c-d052b32c6380","Type":"ContainerDied","Data":"dc64d9036eb9b61d494dbf7d6dee1f340a89e9c9f72316ae6cdbbce7aa276c07"} Oct 07 21:14:40 crc kubenswrapper[4926]: I1007 21:14:40.360334 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerID="100e249606f04d81891685b0722b1d158b92756d193466fcbb473d99a9a79041" exitCode=143 Oct 07 21:14:40 crc kubenswrapper[4926]: I1007 21:14:40.360548 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d546cd4db-lzl8j" event={"ID":"b5d28ce0-4eeb-4408-9f66-d972bc438964","Type":"ContainerDied","Data":"100e249606f04d81891685b0722b1d158b92756d193466fcbb473d99a9a79041"} Oct 07 21:14:40 crc kubenswrapper[4926]: I1007 21:14:40.947479 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:14:40 crc kubenswrapper[4926]: I1007 21:14:40.948947 4926 scope.go:117] "RemoveContainer" containerID="08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9" Oct 07 21:14:40 crc kubenswrapper[4926]: E1007 21:14:40.949309 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:14:41 crc kubenswrapper[4926]: I1007 21:14:41.009046 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-d546cd4db-lzl8j" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.177:9311/healthcheck\": read tcp 10.217.0.2:41338->10.217.0.177:9311: read: connection reset by peer" Oct 07 21:14:41 crc kubenswrapper[4926]: I1007 21:14:41.009273 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-d546cd4db-lzl8j" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.177:9311/healthcheck\": read tcp 10.217.0.2:41330->10.217.0.177:9311: read: connection reset by peer" Oct 07 21:14:41 crc kubenswrapper[4926]: I1007 21:14:41.378223 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerID="87633ac3281fa3d94463173bc2df32fe2e7eec6f4eed36deb19d3f0426f24876" exitCode=0 Oct 07 21:14:41 crc kubenswrapper[4926]: I1007 21:14:41.378231 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d546cd4db-lzl8j" event={"ID":"b5d28ce0-4eeb-4408-9f66-d972bc438964","Type":"ContainerDied","Data":"87633ac3281fa3d94463173bc2df32fe2e7eec6f4eed36deb19d3f0426f24876"} Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.015362 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-55b995649c-qwlp9"] Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.019732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.033181 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.033451 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.034360 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.054421 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-55b995649c-qwlp9"] Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.092676 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28011eca-ae6f-42da-9e6b-27289eb19b2e-log-httpd\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.092968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28011eca-ae6f-42da-9e6b-27289eb19b2e-run-httpd\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.093133 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ddp8\" (UniqueName: \"kubernetes.io/projected/28011eca-ae6f-42da-9e6b-27289eb19b2e-kube-api-access-2ddp8\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.093466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-internal-tls-certs\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.093597 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-config-data\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.093721 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-public-tls-certs\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.093868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28011eca-ae6f-42da-9e6b-27289eb19b2e-etc-swift\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.094015 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-combined-ca-bundle\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-internal-tls-certs\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195282 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-config-data\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195313 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-public-tls-certs\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195351 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28011eca-ae6f-42da-9e6b-27289eb19b2e-etc-swift\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195385 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-combined-ca-bundle\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195416 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28011eca-ae6f-42da-9e6b-27289eb19b2e-log-httpd\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195439 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28011eca-ae6f-42da-9e6b-27289eb19b2e-run-httpd\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.195457 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ddp8\" (UniqueName: \"kubernetes.io/projected/28011eca-ae6f-42da-9e6b-27289eb19b2e-kube-api-access-2ddp8\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.198061 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28011eca-ae6f-42da-9e6b-27289eb19b2e-log-httpd\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.199576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28011eca-ae6f-42da-9e6b-27289eb19b2e-run-httpd\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.202759 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-config-data\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.203089 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-internal-tls-certs\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.203918 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-combined-ca-bundle\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.210959 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ddp8\" (UniqueName: \"kubernetes.io/projected/28011eca-ae6f-42da-9e6b-27289eb19b2e-kube-api-access-2ddp8\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.212646 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28011eca-ae6f-42da-9e6b-27289eb19b2e-public-tls-certs\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.224537 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/28011eca-ae6f-42da-9e6b-27289eb19b2e-etc-swift\") pod \"swift-proxy-55b995649c-qwlp9\" (UID: \"28011eca-ae6f-42da-9e6b-27289eb19b2e\") " pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:42 crc kubenswrapper[4926]: I1007 21:14:42.381016 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.154027 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.154761 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-central-agent" containerID="cri-o://13f3f5c396375822cd81862cf7afb57fe3c0480d24f4b1198b1df0097f952269" gracePeriod=30 Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.154819 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="proxy-httpd" containerID="cri-o://eadcf578f81834730354002bb692272ab0d732f8f30a1a8fc687543ca9404af9" gracePeriod=30 Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.154913 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="sg-core" containerID="cri-o://a457ee3746b3889fd816cdb00c10ab622eeed901dc9247a2f9efcc9cf9e6f2fb" gracePeriod=30 Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.154913 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-notification-agent" containerID="cri-o://e13486028f9002ce9e8c63687bf53bc78ba59386ef5f69d83f0214b2f5cc1c0d" gracePeriod=30 Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.368080 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bf67dc754-22mcb" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.161:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.161:8443: connect: connection refused" Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.368213 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.425361 4926 generic.go:334] "Generic (PLEG): container finished" podID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerID="eadcf578f81834730354002bb692272ab0d732f8f30a1a8fc687543ca9404af9" exitCode=0 Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.425398 4926 generic.go:334] "Generic (PLEG): container finished" podID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerID="a457ee3746b3889fd816cdb00c10ab622eeed901dc9247a2f9efcc9cf9e6f2fb" exitCode=2 Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.425427 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerDied","Data":"eadcf578f81834730354002bb692272ab0d732f8f30a1a8fc687543ca9404af9"} Oct 07 21:14:43 crc kubenswrapper[4926]: I1007 21:14:43.425471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerDied","Data":"a457ee3746b3889fd816cdb00c10ab622eeed901dc9247a2f9efcc9cf9e6f2fb"} Oct 07 21:14:44 crc kubenswrapper[4926]: I1007 21:14:44.453686 4926 generic.go:334] "Generic (PLEG): container finished" podID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerID="e13486028f9002ce9e8c63687bf53bc78ba59386ef5f69d83f0214b2f5cc1c0d" exitCode=0 Oct 07 21:14:44 crc kubenswrapper[4926]: I1007 21:14:44.453931 4926 generic.go:334] "Generic (PLEG): container finished" podID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerID="13f3f5c396375822cd81862cf7afb57fe3c0480d24f4b1198b1df0097f952269" exitCode=0 Oct 07 21:14:44 crc kubenswrapper[4926]: I1007 21:14:44.453882 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerDied","Data":"e13486028f9002ce9e8c63687bf53bc78ba59386ef5f69d83f0214b2f5cc1c0d"} Oct 07 21:14:44 crc kubenswrapper[4926]: I1007 21:14:44.453969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerDied","Data":"13f3f5c396375822cd81862cf7afb57fe3c0480d24f4b1198b1df0097f952269"} Oct 07 21:14:45 crc kubenswrapper[4926]: I1007 21:14:45.596772 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-d546cd4db-lzl8j" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.177:9311/healthcheck\": dial tcp 10.217.0.177:9311: connect: connection refused" Oct 07 21:14:45 crc kubenswrapper[4926]: I1007 21:14:45.596773 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-d546cd4db-lzl8j" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.177:9311/healthcheck\": dial tcp 10.217.0.177:9311: connect: connection refused" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.292436 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bszbk" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.307116 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/920741ff-d4f2-495b-907c-d052b32c6380-etc-machine-id\") pod \"920741ff-d4f2-495b-907c-d052b32c6380\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.307326 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-scripts\") pod \"920741ff-d4f2-495b-907c-d052b32c6380\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.307429 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-combined-ca-bundle\") pod \"920741ff-d4f2-495b-907c-d052b32c6380\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.307478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-db-sync-config-data\") pod \"920741ff-d4f2-495b-907c-d052b32c6380\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.307534 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbrwz\" (UniqueName: \"kubernetes.io/projected/920741ff-d4f2-495b-907c-d052b32c6380-kube-api-access-bbrwz\") pod \"920741ff-d4f2-495b-907c-d052b32c6380\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.307571 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-config-data\") pod \"920741ff-d4f2-495b-907c-d052b32c6380\" (UID: \"920741ff-d4f2-495b-907c-d052b32c6380\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.318332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/920741ff-d4f2-495b-907c-d052b32c6380-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "920741ff-d4f2-495b-907c-d052b32c6380" (UID: "920741ff-d4f2-495b-907c-d052b32c6380"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.319181 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/920741ff-d4f2-495b-907c-d052b32c6380-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.324288 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-scripts" (OuterVolumeSpecName: "scripts") pod "920741ff-d4f2-495b-907c-d052b32c6380" (UID: "920741ff-d4f2-495b-907c-d052b32c6380"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.327397 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "920741ff-d4f2-495b-907c-d052b32c6380" (UID: "920741ff-d4f2-495b-907c-d052b32c6380"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.330544 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920741ff-d4f2-495b-907c-d052b32c6380-kube-api-access-bbrwz" (OuterVolumeSpecName: "kube-api-access-bbrwz") pod "920741ff-d4f2-495b-907c-d052b32c6380" (UID: "920741ff-d4f2-495b-907c-d052b32c6380"). InnerVolumeSpecName "kube-api-access-bbrwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.355431 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "920741ff-d4f2-495b-907c-d052b32c6380" (UID: "920741ff-d4f2-495b-907c-d052b32c6380"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.412098 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-config-data" (OuterVolumeSpecName: "config-data") pod "920741ff-d4f2-495b-907c-d052b32c6380" (UID: "920741ff-d4f2-495b-907c-d052b32c6380"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.421248 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.421284 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbrwz\" (UniqueName: \"kubernetes.io/projected/920741ff-d4f2-495b-907c-d052b32c6380-kube-api-access-bbrwz\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.421296 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.421304 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.421313 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920741ff-d4f2-495b-907c-d052b32c6380-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.484874 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15","Type":"ContainerStarted","Data":"043af7036d2e7b7cc67cdc343c73a4ea67eb82a6573e72708a3ec8e6e9c31aec"} Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.488725 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bszbk" event={"ID":"920741ff-d4f2-495b-907c-d052b32c6380","Type":"ContainerDied","Data":"1a92f22d152f33f7ac2f3e52132f8b42f1eff528e04e5f1124fd264bf95f67db"} Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.488768 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a92f22d152f33f7ac2f3e52132f8b42f1eff528e04e5f1124fd264bf95f67db" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.488838 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bszbk" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.517555 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.9666526260000001 podStartE2EDuration="13.517538049s" podCreationTimestamp="2025-10-07 21:14:33 +0000 UTC" firstStartedPulling="2025-10-07 21:14:34.602120022 +0000 UTC m=+1124.640701172" lastFinishedPulling="2025-10-07 21:14:46.153005445 +0000 UTC m=+1136.191586595" observedRunningTime="2025-10-07 21:14:46.513369489 +0000 UTC m=+1136.551950629" watchObservedRunningTime="2025-10-07 21:14:46.517538049 +0000 UTC m=+1136.556119199" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.566651 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.624220 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6vt9\" (UniqueName: \"kubernetes.io/projected/b5d28ce0-4eeb-4408-9f66-d972bc438964-kube-api-access-t6vt9\") pod \"b5d28ce0-4eeb-4408-9f66-d972bc438964\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.624271 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data\") pod \"b5d28ce0-4eeb-4408-9f66-d972bc438964\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.624355 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5d28ce0-4eeb-4408-9f66-d972bc438964-logs\") pod \"b5d28ce0-4eeb-4408-9f66-d972bc438964\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.624389 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-combined-ca-bundle\") pod \"b5d28ce0-4eeb-4408-9f66-d972bc438964\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.624469 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data-custom\") pod \"b5d28ce0-4eeb-4408-9f66-d972bc438964\" (UID: \"b5d28ce0-4eeb-4408-9f66-d972bc438964\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.625680 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d28ce0-4eeb-4408-9f66-d972bc438964-logs" (OuterVolumeSpecName: "logs") pod "b5d28ce0-4eeb-4408-9f66-d972bc438964" (UID: "b5d28ce0-4eeb-4408-9f66-d972bc438964"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.629742 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d28ce0-4eeb-4408-9f66-d972bc438964-kube-api-access-t6vt9" (OuterVolumeSpecName: "kube-api-access-t6vt9") pod "b5d28ce0-4eeb-4408-9f66-d972bc438964" (UID: "b5d28ce0-4eeb-4408-9f66-d972bc438964"). InnerVolumeSpecName "kube-api-access-t6vt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.632730 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b5d28ce0-4eeb-4408-9f66-d972bc438964" (UID: "b5d28ce0-4eeb-4408-9f66-d972bc438964"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.652603 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.655455 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5d28ce0-4eeb-4408-9f66-d972bc438964" (UID: "b5d28ce0-4eeb-4408-9f66-d972bc438964"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.699277 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data" (OuterVolumeSpecName: "config-data") pod "b5d28ce0-4eeb-4408-9f66-d972bc438964" (UID: "b5d28ce0-4eeb-4408-9f66-d972bc438964"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726170 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-config-data\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726336 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-log-httpd\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726362 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-run-httpd\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726403 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-combined-ca-bundle\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726423 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvp5k\" (UniqueName: \"kubernetes.io/projected/c6ca2cc7-f855-4bbf-aab6-657da7506355-kube-api-access-bvp5k\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726539 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-scripts\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726558 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-sg-core-conf-yaml\") pod \"c6ca2cc7-f855-4bbf-aab6-657da7506355\" (UID: \"c6ca2cc7-f855-4bbf-aab6-657da7506355\") " Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726951 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726982 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5d28ce0-4eeb-4408-9f66-d972bc438964-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.726995 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.727005 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.727015 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6vt9\" (UniqueName: \"kubernetes.io/projected/b5d28ce0-4eeb-4408-9f66-d972bc438964-kube-api-access-t6vt9\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.727023 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5d28ce0-4eeb-4408-9f66-d972bc438964-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.727718 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.735522 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-scripts" (OuterVolumeSpecName: "scripts") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.736259 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6ca2cc7-f855-4bbf-aab6-657da7506355-kube-api-access-bvp5k" (OuterVolumeSpecName: "kube-api-access-bvp5k") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "kube-api-access-bvp5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.765664 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.830036 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.830068 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.830078 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.830088 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6ca2cc7-f855-4bbf-aab6-657da7506355-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.830097 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvp5k\" (UniqueName: \"kubernetes.io/projected/c6ca2cc7-f855-4bbf-aab6-657da7506355-kube-api-access-bvp5k\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.853638 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.906671 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-55b995649c-qwlp9"] Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.934542 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.952453 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.952732 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-log" containerID="cri-o://3f91018b9c9a7df718757a1a4f896e3345bb3fad2f99da4b8ad99a66b08ea24a" gracePeriod=30 Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.953139 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-httpd" containerID="cri-o://0cc6049dec82fff93de6a1cff120ab74b4312141366363d17ead23276dfcc5dd" gracePeriod=30 Oct 07 21:14:46 crc kubenswrapper[4926]: I1007 21:14:46.990357 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-config-data" (OuterVolumeSpecName: "config-data") pod "c6ca2cc7-f855-4bbf-aab6-657da7506355" (UID: "c6ca2cc7-f855-4bbf-aab6-657da7506355"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.036072 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ca2cc7-f855-4bbf-aab6-657da7506355-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.531519 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6ca2cc7-f855-4bbf-aab6-657da7506355","Type":"ContainerDied","Data":"e3fbccb17bd4c2092b32d5255dd04a4496efd72ba175fc4b6ef5689353b5b269"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.532065 4926 scope.go:117] "RemoveContainer" containerID="eadcf578f81834730354002bb692272ab0d732f8f30a1a8fc687543ca9404af9" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.532427 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.551712 4926 generic.go:334] "Generic (PLEG): container finished" podID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerID="c998b9bb106c97402a01e919798bd57bd94c1b298710141166203f882e278cb6" exitCode=137 Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.551804 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bf67dc754-22mcb" event={"ID":"3739cb7c-d23c-462e-a7ec-0bd07c2a0149","Type":"ContainerDied","Data":"c998b9bb106c97402a01e919798bd57bd94c1b298710141166203f882e278cb6"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.551839 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bf67dc754-22mcb" event={"ID":"3739cb7c-d23c-462e-a7ec-0bd07c2a0149","Type":"ContainerDied","Data":"4c161a81298b890f572420a6a12514907b511623530dac51dacf294648350900"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.551855 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c161a81298b890f572420a6a12514907b511623530dac51dacf294648350900" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.564991 4926 generic.go:334] "Generic (PLEG): container finished" podID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerID="3f91018b9c9a7df718757a1a4f896e3345bb3fad2f99da4b8ad99a66b08ea24a" exitCode=143 Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.565345 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e47c8f35-5fce-4dc5-b86d-895606525d5f","Type":"ContainerDied","Data":"3f91018b9c9a7df718757a1a4f896e3345bb3fad2f99da4b8ad99a66b08ea24a"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.567305 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-55b995649c-qwlp9" event={"ID":"28011eca-ae6f-42da-9e6b-27289eb19b2e","Type":"ContainerStarted","Data":"a09352612f3233bc200f9fd600e03a59ba42912f47c070759f1cc692ba6436b3"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.567420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-55b995649c-qwlp9" event={"ID":"28011eca-ae6f-42da-9e6b-27289eb19b2e","Type":"ContainerStarted","Data":"badcd453e03cc8cfe14a762017137b3c630534a11db4e31eca4cc21f97b4b5a7"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.584690 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.586802 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d546cd4db-lzl8j" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.590909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d546cd4db-lzl8j" event={"ID":"b5d28ce0-4eeb-4408-9f66-d972bc438964","Type":"ContainerDied","Data":"7dd70d15901324771575244211e64287b6e626872324527023a632759b2992c7"} Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.613572 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.625776 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.653920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-combined-ca-bundle\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.654039 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-tls-certs\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.654081 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-config-data\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.654173 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-logs\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.654249 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-secret-key\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.654317 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-scripts\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.654368 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npwj9\" (UniqueName: \"kubernetes.io/projected/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-kube-api-access-npwj9\") pod \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\" (UID: \"3739cb7c-d23c-462e-a7ec-0bd07c2a0149\") " Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.659058 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.659933 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.659949 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.659965 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-notification-agent" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.659972 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-notification-agent" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.659990 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="proxy-httpd" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.659997 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="proxy-httpd" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.660012 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660018 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.660029 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-central-agent" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660035 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-central-agent" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.660049 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="sg-core" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660056 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="sg-core" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.660079 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920741ff-d4f2-495b-907c-d052b32c6380" containerName="cinder-db-sync" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660085 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="920741ff-d4f2-495b-907c-d052b32c6380" containerName="cinder-db-sync" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.660101 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api-log" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660107 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api-log" Oct 07 21:14:47 crc kubenswrapper[4926]: E1007 21:14:47.660115 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon-log" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660121 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon-log" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660306 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-notification-agent" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660319 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="sg-core" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660333 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="920741ff-d4f2-495b-907c-d052b32c6380" containerName="cinder-db-sync" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660348 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api-log" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660362 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" containerName="barbican-api" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660371 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660393 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="proxy-httpd" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660407 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" containerName="horizon-log" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.660415 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" containerName="ceilometer-central-agent" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.670346 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-logs" (OuterVolumeSpecName: "logs") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.698273 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.699738 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.707520 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-kube-api-access-npwj9" (OuterVolumeSpecName: "kube-api-access-npwj9") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "kube-api-access-npwj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.714213 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.714430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.714540 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.728959 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.720004 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hw4jl" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.735429 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.740492 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.748298 4926 scope.go:117] "RemoveContainer" containerID="a457ee3746b3889fd816cdb00c10ab622eeed901dc9247a2f9efcc9cf9e6f2fb" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.756920 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-config-data\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.756966 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.757061 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dca9ab35-65a2-4034-921a-a295cde7e581-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.757082 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-run-httpd\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.757109 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.757133 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.760701 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.761038 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.757185 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp6zh\" (UniqueName: \"kubernetes.io/projected/3ef93fed-2339-4404-adcb-76fcb021c782-kube-api-access-cp6zh\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.763701 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-log-httpd\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.763739 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-scripts\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.763840 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-scripts\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.763911 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntkgd\" (UniqueName: \"kubernetes.io/projected/dca9ab35-65a2-4034-921a-a295cde7e581-kube-api-access-ntkgd\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.764022 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.764101 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.764284 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.764300 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.764315 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npwj9\" (UniqueName: \"kubernetes.io/projected/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-kube-api-access-npwj9\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.764326 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.768668 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-scripts" (OuterVolumeSpecName: "scripts") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.781467 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.799351 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.804241 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-d546cd4db-lzl8j"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.809610 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-config-data" (OuterVolumeSpecName: "config-data") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.818952 4926 scope.go:117] "RemoveContainer" containerID="e13486028f9002ce9e8c63687bf53bc78ba59386ef5f69d83f0214b2f5cc1c0d" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.819756 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-d546cd4db-lzl8j"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.852925 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3739cb7c-d23c-462e-a7ec-0bd07c2a0149" (UID: "3739cb7c-d23c-462e-a7ec-0bd07c2a0149"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866357 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntkgd\" (UniqueName: \"kubernetes.io/projected/dca9ab35-65a2-4034-921a-a295cde7e581-kube-api-access-ntkgd\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866418 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866451 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866495 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-config-data\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866513 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866578 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dca9ab35-65a2-4034-921a-a295cde7e581-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866598 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-run-httpd\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866640 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866670 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp6zh\" (UniqueName: \"kubernetes.io/projected/3ef93fed-2339-4404-adcb-76fcb021c782-kube-api-access-cp6zh\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866699 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-log-httpd\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866718 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-scripts\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866748 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-scripts\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866809 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866822 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.866831 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3739cb7c-d23c-462e-a7ec-0bd07c2a0149-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.867676 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-run-httpd\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.869323 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84c5b587c5-jvms2"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.875533 4926 scope.go:117] "RemoveContainer" containerID="13f3f5c396375822cd81862cf7afb57fe3c0480d24f4b1198b1df0097f952269" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.876207 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dca9ab35-65a2-4034-921a-a295cde7e581-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.876466 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-log-httpd\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.876654 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-config-data\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.877285 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.878628 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.876137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.881033 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.882743 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.882913 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.897181 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntkgd\" (UniqueName: \"kubernetes.io/projected/dca9ab35-65a2-4034-921a-a295cde7e581-kube-api-access-ntkgd\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.903838 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-scripts\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.904445 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-scripts\") pod \"cinder-scheduler-0\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " pod="openstack/cinder-scheduler-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.908216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp6zh\" (UniqueName: \"kubernetes.io/projected/3ef93fed-2339-4404-adcb-76fcb021c782-kube-api-access-cp6zh\") pod \"ceilometer-0\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " pod="openstack/ceilometer-0" Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.930652 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84c5b587c5-jvms2"] Oct 07 21:14:47 crc kubenswrapper[4926]: I1007 21:14:47.945565 4926 scope.go:117] "RemoveContainer" containerID="87633ac3281fa3d94463173bc2df32fe2e7eec6f4eed36deb19d3f0426f24876" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.004402 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-nb\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.004497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-sb\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.004546 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-swift-storage-0\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.004809 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-svc\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.004925 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d872c\" (UniqueName: \"kubernetes.io/projected/a772eddc-1462-4636-a4ba-1340b24dc102-kube-api-access-d872c\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.004974 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-config\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.074030 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.116578 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-nb\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.116727 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-sb\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.116766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-swift-storage-0\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.116901 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-svc\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.116979 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d872c\" (UniqueName: \"kubernetes.io/projected/a772eddc-1462-4636-a4ba-1340b24dc102-kube-api-access-d872c\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.117013 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-config\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.125308 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-config\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.125417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-svc\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.126693 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-swift-storage-0\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.127410 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-sb\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.128703 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.135949 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-nb\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.142429 4926 scope.go:117] "RemoveContainer" containerID="100e249606f04d81891685b0722b1d158b92756d193466fcbb473d99a9a79041" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.149965 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d872c\" (UniqueName: \"kubernetes.io/projected/a772eddc-1462-4636-a4ba-1340b24dc102-kube-api-access-d872c\") pod \"dnsmasq-dns-84c5b587c5-jvms2\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.193497 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.195302 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.203522 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.208064 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.248274 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320498 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320594 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bdab5b7-41be-4087-b654-fd2434a04aaa-logs\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320633 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-scripts\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320652 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq6ph\" (UniqueName: \"kubernetes.io/projected/3bdab5b7-41be-4087-b654-fd2434a04aaa-kube-api-access-zq6ph\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320778 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data-custom\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.320821 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bdab5b7-41be-4087-b654-fd2434a04aaa-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.424859 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bdab5b7-41be-4087-b654-fd2434a04aaa-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425285 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bdab5b7-41be-4087-b654-fd2434a04aaa-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425400 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bdab5b7-41be-4087-b654-fd2434a04aaa-logs\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425436 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-scripts\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425458 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425572 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq6ph\" (UniqueName: \"kubernetes.io/projected/3bdab5b7-41be-4087-b654-fd2434a04aaa-kube-api-access-zq6ph\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425601 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data-custom\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.425891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bdab5b7-41be-4087-b654-fd2434a04aaa-logs\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.433153 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-scripts\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.434457 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.434768 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.447347 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data-custom\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.451449 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq6ph\" (UniqueName: \"kubernetes.io/projected/3bdab5b7-41be-4087-b654-fd2434a04aaa-kube-api-access-zq6ph\") pod \"cinder-api-0\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.605937 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.607518 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-55b995649c-qwlp9" event={"ID":"28011eca-ae6f-42da-9e6b-27289eb19b2e","Type":"ContainerStarted","Data":"94106a5d352cb0e0fba484ff1db15f00b795ca8604374ab2fad6305071e41b86"} Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.607588 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.607609 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.616504 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bf67dc754-22mcb" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.653462 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-55b995649c-qwlp9" podStartSLOduration=7.653433415 podStartE2EDuration="7.653433415s" podCreationTimestamp="2025-10-07 21:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:48.63212641 +0000 UTC m=+1138.670707560" watchObservedRunningTime="2025-10-07 21:14:48.653433415 +0000 UTC m=+1138.692014565" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.678249 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bf67dc754-22mcb"] Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.702636 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d28ce0-4eeb-4408-9f66-d972bc438964" path="/var/lib/kubelet/pods/b5d28ce0-4eeb-4408-9f66-d972bc438964/volumes" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.703247 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6ca2cc7-f855-4bbf-aab6-657da7506355" path="/var/lib/kubelet/pods/c6ca2cc7-f855-4bbf-aab6-657da7506355/volumes" Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.703984 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bf67dc754-22mcb"] Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.785878 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.897588 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84c5b587c5-jvms2"] Oct 07 21:14:48 crc kubenswrapper[4926]: I1007 21:14:48.908530 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.045523 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-d299q"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.046945 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.058762 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-d299q"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.142496 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnff5\" (UniqueName: \"kubernetes.io/projected/8039b214-d022-4525-9ae6-c7b11921aa82-kube-api-access-gnff5\") pod \"nova-api-db-create-d299q\" (UID: \"8039b214-d022-4525-9ae6-c7b11921aa82\") " pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.211697 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-8xkbk"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.213133 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.231017 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8xkbk"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.246479 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnff5\" (UniqueName: \"kubernetes.io/projected/8039b214-d022-4525-9ae6-c7b11921aa82-kube-api-access-gnff5\") pod \"nova-api-db-create-d299q\" (UID: \"8039b214-d022-4525-9ae6-c7b11921aa82\") " pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.246659 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2mjz\" (UniqueName: \"kubernetes.io/projected/5c212508-2303-4aac-86eb-0c5ee339e517-kube-api-access-b2mjz\") pod \"nova-cell0-db-create-8xkbk\" (UID: \"5c212508-2303-4aac-86eb-0c5ee339e517\") " pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.261794 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.262184 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-log" containerID="cri-o://4fe99c09c2dea580c5940654bd065d473abfef0c0aba98c752107b97fd028994" gracePeriod=30 Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.262663 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-httpd" containerID="cri-o://0e597c2d8da33aa0eab78483b2abf155682e10378c17a70a1f82b364957e5112" gracePeriod=30 Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.276530 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnff5\" (UniqueName: \"kubernetes.io/projected/8039b214-d022-4525-9ae6-c7b11921aa82-kube-api-access-gnff5\") pod \"nova-api-db-create-d299q\" (UID: \"8039b214-d022-4525-9ae6-c7b11921aa82\") " pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.349188 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2mjz\" (UniqueName: \"kubernetes.io/projected/5c212508-2303-4aac-86eb-0c5ee339e517-kube-api-access-b2mjz\") pod \"nova-cell0-db-create-8xkbk\" (UID: \"5c212508-2303-4aac-86eb-0c5ee339e517\") " pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.352208 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.352952 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.361689 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-86p2f"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.363382 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.426442 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2mjz\" (UniqueName: \"kubernetes.io/projected/5c212508-2303-4aac-86eb-0c5ee339e517-kube-api-access-b2mjz\") pod \"nova-cell0-db-create-8xkbk\" (UID: \"5c212508-2303-4aac-86eb-0c5ee339e517\") " pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.448714 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-86p2f"] Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.451745 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bllfl\" (UniqueName: \"kubernetes.io/projected/db0e9ee6-03fb-45da-abda-e9af50b92dc7-kube-api-access-bllfl\") pod \"nova-cell1-db-create-86p2f\" (UID: \"db0e9ee6-03fb-45da-abda-e9af50b92dc7\") " pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.562369 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bllfl\" (UniqueName: \"kubernetes.io/projected/db0e9ee6-03fb-45da-abda-e9af50b92dc7-kube-api-access-bllfl\") pod \"nova-cell1-db-create-86p2f\" (UID: \"db0e9ee6-03fb-45da-abda-e9af50b92dc7\") " pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.587690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bllfl\" (UniqueName: \"kubernetes.io/projected/db0e9ee6-03fb-45da-abda-e9af50b92dc7-kube-api-access-bllfl\") pod \"nova-cell1-db-create-86p2f\" (UID: \"db0e9ee6-03fb-45da-abda-e9af50b92dc7\") " pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.632740 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3bdab5b7-41be-4087-b654-fd2434a04aaa","Type":"ContainerStarted","Data":"60d9c7c577cc7229bc0d0d987ad45dfc60f47dda6eaf86edb80adec9efeb1faf"} Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.636547 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dca9ab35-65a2-4034-921a-a295cde7e581","Type":"ContainerStarted","Data":"ccb840e6ebdc48bcdc07f9bb76847a6138f6773c08de846ae78279ce55616146"} Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.647300 4926 generic.go:334] "Generic (PLEG): container finished" podID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerID="4fe99c09c2dea580c5940654bd065d473abfef0c0aba98c752107b97fd028994" exitCode=143 Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.647377 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1598a03a-83d6-45b1-830f-5ef8c37f177a","Type":"ContainerDied","Data":"4fe99c09c2dea580c5940654bd065d473abfef0c0aba98c752107b97fd028994"} Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.650092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" event={"ID":"a772eddc-1462-4636-a4ba-1340b24dc102","Type":"ContainerStarted","Data":"572b636518e49533f8f9268c82d2d59424c52ca59da24ce9029fc7b6c2fcde86"} Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.661965 4926 generic.go:334] "Generic (PLEG): container finished" podID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerID="0cc6049dec82fff93de6a1cff120ab74b4312141366363d17ead23276dfcc5dd" exitCode=0 Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.662085 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e47c8f35-5fce-4dc5-b86d-895606525d5f","Type":"ContainerDied","Data":"0cc6049dec82fff93de6a1cff120ab74b4312141366363d17ead23276dfcc5dd"} Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.674051 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerStarted","Data":"a96ceed949664e3e257628a0dcba5dfd61b3fa7ab6ac5efb8b137cc36f4254ab"} Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.711876 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.751529 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:49 crc kubenswrapper[4926]: I1007 21:14:49.881251 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.044973 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-public-tls-certs\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045024 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tblnt\" (UniqueName: \"kubernetes.io/projected/e47c8f35-5fce-4dc5-b86d-895606525d5f-kube-api-access-tblnt\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045065 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-config-data\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045092 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-scripts\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045163 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-logs\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045227 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-combined-ca-bundle\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045264 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-httpd-run\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.045293 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"e47c8f35-5fce-4dc5-b86d-895606525d5f\" (UID: \"e47c8f35-5fce-4dc5-b86d-895606525d5f\") " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.048755 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-logs" (OuterVolumeSpecName: "logs") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.048914 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.059373 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.076387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-scripts" (OuterVolumeSpecName: "scripts") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.085498 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e47c8f35-5fce-4dc5-b86d-895606525d5f-kube-api-access-tblnt" (OuterVolumeSpecName: "kube-api-access-tblnt") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "kube-api-access-tblnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.171220 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tblnt\" (UniqueName: \"kubernetes.io/projected/e47c8f35-5fce-4dc5-b86d-895606525d5f-kube-api-access-tblnt\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.171266 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.171281 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.171291 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e47c8f35-5fce-4dc5-b86d-895606525d5f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.171324 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.212326 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.234547 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-d299q"] Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.258382 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-config-data" (OuterVolumeSpecName: "config-data") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.272879 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.272911 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.304241 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e47c8f35-5fce-4dc5-b86d-895606525d5f" (UID: "e47c8f35-5fce-4dc5-b86d-895606525d5f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.349495 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.379484 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.379513 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e47c8f35-5fce-4dc5-b86d-895606525d5f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.565828 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:50 crc kubenswrapper[4926]: W1007 21:14:50.653989 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8039b214_d022_4525_9ae6_c7b11921aa82.slice/crio-825ed6858763b3cf8564e35b1bebf521a8a2c1d20214fdf09b1b111e4df94529 WatchSource:0}: Error finding container 825ed6858763b3cf8564e35b1bebf521a8a2c1d20214fdf09b1b111e4df94529: Status 404 returned error can't find the container with id 825ed6858763b3cf8564e35b1bebf521a8a2c1d20214fdf09b1b111e4df94529 Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.836826 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3739cb7c-d23c-462e-a7ec-0bd07c2a0149" path="/var/lib/kubelet/pods/3739cb7c-d23c-462e-a7ec-0bd07c2a0149/volumes" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.837726 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.837753 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerStarted","Data":"c2978894f94e376ff151ea0b1c8a3968b11fa5960e7311b617c92b908f8e8925"} Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.854530 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d299q" event={"ID":"8039b214-d022-4525-9ae6-c7b11921aa82","Type":"ContainerStarted","Data":"825ed6858763b3cf8564e35b1bebf521a8a2c1d20214fdf09b1b111e4df94529"} Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.856137 4926 generic.go:334] "Generic (PLEG): container finished" podID="a772eddc-1462-4636-a4ba-1340b24dc102" containerID="e28592f2a9801ea3408b48216016b5e85e9c0138194b9419c3c0fa4eba009d6f" exitCode=0 Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.856213 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" event={"ID":"a772eddc-1462-4636-a4ba-1340b24dc102","Type":"ContainerDied","Data":"e28592f2a9801ea3408b48216016b5e85e9c0138194b9419c3c0fa4eba009d6f"} Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.868564 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e47c8f35-5fce-4dc5-b86d-895606525d5f","Type":"ContainerDied","Data":"37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69"} Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.868622 4926 scope.go:117] "RemoveContainer" containerID="0cc6049dec82fff93de6a1cff120ab74b4312141366363d17ead23276dfcc5dd" Oct 07 21:14:50 crc kubenswrapper[4926]: I1007 21:14:50.868807 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.049486 4926 scope.go:117] "RemoveContainer" containerID="3f91018b9c9a7df718757a1a4f896e3345bb3fad2f99da4b8ad99a66b08ea24a" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.069828 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-86p2f"] Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.162058 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.193897 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.167:9292/healthcheck\": read tcp 10.217.0.2:36510->10.217.0.167:9292: read: connection reset by peer" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.194258 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.167:9292/healthcheck\": read tcp 10.217.0.2:36520->10.217.0.167:9292: read: connection reset by peer" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.205555 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8xkbk"] Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.222606 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.223640 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:14:51 crc kubenswrapper[4926]: E1007 21:14:51.224047 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-httpd" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.224060 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-httpd" Oct 07 21:14:51 crc kubenswrapper[4926]: E1007 21:14:51.224072 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-log" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.224079 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-log" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.224278 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-httpd" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.224300 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" containerName="glance-log" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.225298 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.238352 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.238993 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.277005 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.312856 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.312917 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15c1a57a-dd1c-4955-ab28-86f0d09faffb-logs\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.312949 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15c1a57a-dd1c-4955-ab28-86f0d09faffb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.312981 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.313004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-config-data\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.313025 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-scripts\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.313042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.313090 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55vds\" (UniqueName: \"kubernetes.io/projected/15c1a57a-dd1c-4955-ab28-86f0d09faffb-kube-api-access-55vds\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.414780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415174 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-config-data\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415220 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-scripts\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415245 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415315 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55vds\" (UniqueName: \"kubernetes.io/projected/15c1a57a-dd1c-4955-ab28-86f0d09faffb-kube-api-access-55vds\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415391 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415424 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15c1a57a-dd1c-4955-ab28-86f0d09faffb-logs\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.415467 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15c1a57a-dd1c-4955-ab28-86f0d09faffb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.416172 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15c1a57a-dd1c-4955-ab28-86f0d09faffb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.416459 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.416826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15c1a57a-dd1c-4955-ab28-86f0d09faffb-logs\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.425459 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-scripts\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.426124 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.428017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.429507 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c1a57a-dd1c-4955-ab28-86f0d09faffb-config-data\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.436011 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55vds\" (UniqueName: \"kubernetes.io/projected/15c1a57a-dd1c-4955-ab28-86f0d09faffb-kube-api-access-55vds\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.525583 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"15c1a57a-dd1c-4955-ab28-86f0d09faffb\") " pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.600894 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 21:14:51 crc kubenswrapper[4926]: I1007 21:14:51.987988 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerStarted","Data":"682ec989e0923a62b8b1d4ccbe15cbce926d45bc56fc13bf4a977837b79dbec7"} Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.030228 4926 generic.go:334] "Generic (PLEG): container finished" podID="8039b214-d022-4525-9ae6-c7b11921aa82" containerID="029f6ea5a5085e74a95d1d3e5149a8ecc6a141bd0908624a8da52cf40222fa99" exitCode=0 Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.030294 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d299q" event={"ID":"8039b214-d022-4525-9ae6-c7b11921aa82","Type":"ContainerDied","Data":"029f6ea5a5085e74a95d1d3e5149a8ecc6a141bd0908624a8da52cf40222fa99"} Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.039102 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8xkbk" event={"ID":"5c212508-2303-4aac-86eb-0c5ee339e517","Type":"ContainerStarted","Data":"d2a532f9e15546fd38f8ba9aa0e8c6479db0cdfe5d251cbd5d3ecd76e2537e38"} Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.040395 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dca9ab35-65a2-4034-921a-a295cde7e581","Type":"ContainerStarted","Data":"b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71"} Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.050615 4926 generic.go:334] "Generic (PLEG): container finished" podID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerID="0e597c2d8da33aa0eab78483b2abf155682e10378c17a70a1f82b364957e5112" exitCode=0 Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.050881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1598a03a-83d6-45b1-830f-5ef8c37f177a","Type":"ContainerDied","Data":"0e597c2d8da33aa0eab78483b2abf155682e10378c17a70a1f82b364957e5112"} Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.053486 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-86p2f" event={"ID":"db0e9ee6-03fb-45da-abda-e9af50b92dc7","Type":"ContainerStarted","Data":"d724aeee9b82ad493b13697cc0d6fc886f555e52529475d3d83ba6b8f6e12f19"} Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.280994 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.334545 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.334951 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-logs\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.335014 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-internal-tls-certs\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.335125 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-combined-ca-bundle\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.335269 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-config-data\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.335312 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-scripts\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.335344 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-httpd-run\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.335373 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhnpf\" (UniqueName: \"kubernetes.io/projected/1598a03a-83d6-45b1-830f-5ef8c37f177a-kube-api-access-xhnpf\") pod \"1598a03a-83d6-45b1-830f-5ef8c37f177a\" (UID: \"1598a03a-83d6-45b1-830f-5ef8c37f177a\") " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.339690 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-logs" (OuterVolumeSpecName: "logs") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.340042 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.345076 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1598a03a-83d6-45b1-830f-5ef8c37f177a-kube-api-access-xhnpf" (OuterVolumeSpecName: "kube-api-access-xhnpf") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "kube-api-access-xhnpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.350490 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.351383 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-scripts" (OuterVolumeSpecName: "scripts") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.417807 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.437663 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.437701 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.437711 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhnpf\" (UniqueName: \"kubernetes.io/projected/1598a03a-83d6-45b1-830f-5ef8c37f177a-kube-api-access-xhnpf\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.437731 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.437742 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1598a03a-83d6-45b1-830f-5ef8c37f177a-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.557255 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.565257 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.587767 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.625148 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.645690 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.645729 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.664568 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-config-data" (OuterVolumeSpecName: "config-data") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.688044 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1598a03a-83d6-45b1-830f-5ef8c37f177a" (UID: "1598a03a-83d6-45b1-830f-5ef8c37f177a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.704466 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e47c8f35-5fce-4dc5-b86d-895606525d5f" path="/var/lib/kubelet/pods/e47c8f35-5fce-4dc5-b86d-895606525d5f/volumes" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.747890 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:52 crc kubenswrapper[4926]: I1007 21:14:52.747928 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598a03a-83d6-45b1-830f-5ef8c37f177a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.057510 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-b64c68dfc-k895v" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.187813 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5df5c749bd-sp7b6"] Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.188451 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5df5c749bd-sp7b6" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-httpd" containerID="cri-o://eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31" gracePeriod=30 Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.201128 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5df5c749bd-sp7b6" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-api" containerID="cri-o://f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a" gracePeriod=30 Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.243003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerStarted","Data":"c97a1128966df6323a30edc576ce4e36b29dd5694c631091b9201d3843aad377"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.273121 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3bdab5b7-41be-4087-b654-fd2434a04aaa","Type":"ContainerStarted","Data":"f6578f217c33ea24cb13af9db1810f8190e7a8bdcfe37c0c1cd2982992eb3f4f"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.285223 4926 generic.go:334] "Generic (PLEG): container finished" podID="5c212508-2303-4aac-86eb-0c5ee339e517" containerID="5f1fbd0881bdd18749cb9412158fc75fa28e096f8839d0aac8c07df15707a678" exitCode=0 Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.285321 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8xkbk" event={"ID":"5c212508-2303-4aac-86eb-0c5ee339e517","Type":"ContainerDied","Data":"5f1fbd0881bdd18749cb9412158fc75fa28e096f8839d0aac8c07df15707a678"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.291145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dca9ab35-65a2-4034-921a-a295cde7e581","Type":"ContainerStarted","Data":"ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.320177 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.323282 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1598a03a-83d6-45b1-830f-5ef8c37f177a","Type":"ContainerDied","Data":"0db1a2000501bde8fe124dff5e1449fb8311d8af48b5be77623e560f70e902f0"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.323362 4926 scope.go:117] "RemoveContainer" containerID="0e597c2d8da33aa0eab78483b2abf155682e10378c17a70a1f82b364957e5112" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.336332 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"15c1a57a-dd1c-4955-ab28-86f0d09faffb","Type":"ContainerStarted","Data":"a0aa7ef56109dcb195aae2d854f139424d0a326b71df3736ce70ac2071770c13"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.360052 4926 generic.go:334] "Generic (PLEG): container finished" podID="db0e9ee6-03fb-45da-abda-e9af50b92dc7" containerID="90c6e3d92954ca83fae6c6630608d9d34810f761d46a92c55da6834b453a6897" exitCode=0 Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.360115 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-86p2f" event={"ID":"db0e9ee6-03fb-45da-abda-e9af50b92dc7","Type":"ContainerDied","Data":"90c6e3d92954ca83fae6c6630608d9d34810f761d46a92c55da6834b453a6897"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.363742 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" event={"ID":"a772eddc-1462-4636-a4ba-1340b24dc102","Type":"ContainerStarted","Data":"ad0f1088b7fafad316bdcc107a5ed82e953a65ca275ebaf97ec7db01a4518852"} Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.363773 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.368452 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.857917526 podStartE2EDuration="6.36842104s" podCreationTimestamp="2025-10-07 21:14:47 +0000 UTC" firstStartedPulling="2025-10-07 21:14:48.776279238 +0000 UTC m=+1138.814860388" lastFinishedPulling="2025-10-07 21:14:49.286782752 +0000 UTC m=+1139.325363902" observedRunningTime="2025-10-07 21:14:53.332769741 +0000 UTC m=+1143.371350891" watchObservedRunningTime="2025-10-07 21:14:53.36842104 +0000 UTC m=+1143.407002190" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.412277 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.422417 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.436547 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:14:53 crc kubenswrapper[4926]: E1007 21:14:53.437048 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-log" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.437064 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-log" Oct 07 21:14:53 crc kubenswrapper[4926]: E1007 21:14:53.437088 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-httpd" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.437093 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-httpd" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.437322 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-httpd" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.437341 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" containerName="glance-log" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.441300 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" podStartSLOduration=6.4412702809999995 podStartE2EDuration="6.441270281s" podCreationTimestamp="2025-10-07 21:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:53.439735426 +0000 UTC m=+1143.478316586" watchObservedRunningTime="2025-10-07 21:14:53.441270281 +0000 UTC m=+1143.479851421" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.445376 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.451435 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.451979 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.483220 4926 scope.go:117] "RemoveContainer" containerID="4fe99c09c2dea580c5940654bd065d473abfef0c0aba98c752107b97fd028994" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.492757 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.492974 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.493030 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-logs\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.493116 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.493141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-268ps\" (UniqueName: \"kubernetes.io/projected/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-kube-api-access-268ps\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.493220 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.493251 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.493299 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.531623 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.594917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.594958 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-logs\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595019 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595040 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-268ps\" (UniqueName: \"kubernetes.io/projected/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-kube-api-access-268ps\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595496 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595526 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595564 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595629 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-logs\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.596352 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.595553 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.603372 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.606850 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.611834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.616519 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.626956 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-268ps\" (UniqueName: \"kubernetes.io/projected/e4b6a533-2f48-428f-b735-3ceac2e2d7c7-kube-api-access-268ps\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.672069 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e4b6a533-2f48-428f-b735-3ceac2e2d7c7\") " pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.774339 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 21:14:53 crc kubenswrapper[4926]: I1007 21:14:53.902409 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.006293 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnff5\" (UniqueName: \"kubernetes.io/projected/8039b214-d022-4525-9ae6-c7b11921aa82-kube-api-access-gnff5\") pod \"8039b214-d022-4525-9ae6-c7b11921aa82\" (UID: \"8039b214-d022-4525-9ae6-c7b11921aa82\") " Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.019847 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8039b214-d022-4525-9ae6-c7b11921aa82-kube-api-access-gnff5" (OuterVolumeSpecName: "kube-api-access-gnff5") pod "8039b214-d022-4525-9ae6-c7b11921aa82" (UID: "8039b214-d022-4525-9ae6-c7b11921aa82"). InnerVolumeSpecName "kube-api-access-gnff5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.110607 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnff5\" (UniqueName: \"kubernetes.io/projected/8039b214-d022-4525-9ae6-c7b11921aa82-kube-api-access-gnff5\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.376503 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3bdab5b7-41be-4087-b654-fd2434a04aaa","Type":"ContainerStarted","Data":"6e8c53be911ace17038a2e6cadd015035402ab7fdd4bc20dffa0eabf1a87b53e"} Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.376668 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api-log" containerID="cri-o://f6578f217c33ea24cb13af9db1810f8190e7a8bdcfe37c0c1cd2982992eb3f4f" gracePeriod=30 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.376935 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.377320 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api" containerID="cri-o://6e8c53be911ace17038a2e6cadd015035402ab7fdd4bc20dffa0eabf1a87b53e" gracePeriod=30 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.401464 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"15c1a57a-dd1c-4955-ab28-86f0d09faffb","Type":"ContainerStarted","Data":"50c3777345539dbe23067202a1f0d1fb956391d06e52650d26245848c1db1f22"} Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.410364 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerStarted","Data":"19b4e58f9466c1b8227f113d5aacfb990e8ca33333ceaf1283f31330a9dd1895"} Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.410540 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-central-agent" containerID="cri-o://c2978894f94e376ff151ea0b1c8a3968b11fa5960e7311b617c92b908f8e8925" gracePeriod=30 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.410786 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.411076 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="proxy-httpd" containerID="cri-o://19b4e58f9466c1b8227f113d5aacfb990e8ca33333ceaf1283f31330a9dd1895" gracePeriod=30 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.411128 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="sg-core" containerID="cri-o://c97a1128966df6323a30edc576ce4e36b29dd5694c631091b9201d3843aad377" gracePeriod=30 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.411169 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-notification-agent" containerID="cri-o://682ec989e0923a62b8b1d4ccbe15cbce926d45bc56fc13bf4a977837b79dbec7" gracePeriod=30 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.412523 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.412491253 podStartE2EDuration="7.412491253s" podCreationTimestamp="2025-10-07 21:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:54.391971871 +0000 UTC m=+1144.430553021" watchObservedRunningTime="2025-10-07 21:14:54.412491253 +0000 UTC m=+1144.451072403" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.424304 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d299q" event={"ID":"8039b214-d022-4525-9ae6-c7b11921aa82","Type":"ContainerDied","Data":"825ed6858763b3cf8564e35b1bebf521a8a2c1d20214fdf09b1b111e4df94529"} Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.424771 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="825ed6858763b3cf8564e35b1bebf521a8a2c1d20214fdf09b1b111e4df94529" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.424852 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d299q" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.436298 4926 generic.go:334] "Generic (PLEG): container finished" podID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerID="eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31" exitCode=0 Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.436491 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df5c749bd-sp7b6" event={"ID":"e4537dd0-c22a-4879-9fa2-d02f3403172d","Type":"ContainerDied","Data":"eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31"} Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.474720 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.737685708 podStartE2EDuration="7.474698317s" podCreationTimestamp="2025-10-07 21:14:47 +0000 UTC" firstStartedPulling="2025-10-07 21:14:49.145021094 +0000 UTC m=+1139.183602244" lastFinishedPulling="2025-10-07 21:14:53.882033703 +0000 UTC m=+1143.920614853" observedRunningTime="2025-10-07 21:14:54.457724608 +0000 UTC m=+1144.496305758" watchObservedRunningTime="2025-10-07 21:14:54.474698317 +0000 UTC m=+1144.513279467" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.504919 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.682748 4926 scope.go:117] "RemoveContainer" containerID="08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.733893 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1598a03a-83d6-45b1-830f-5ef8c37f177a" path="/var/lib/kubelet/pods/1598a03a-83d6-45b1-830f-5ef8c37f177a/volumes" Oct 07 21:14:54 crc kubenswrapper[4926]: I1007 21:14:54.977418 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.044724 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bllfl\" (UniqueName: \"kubernetes.io/projected/db0e9ee6-03fb-45da-abda-e9af50b92dc7-kube-api-access-bllfl\") pod \"db0e9ee6-03fb-45da-abda-e9af50b92dc7\" (UID: \"db0e9ee6-03fb-45da-abda-e9af50b92dc7\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.076588 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0e9ee6-03fb-45da-abda-e9af50b92dc7-kube-api-access-bllfl" (OuterVolumeSpecName: "kube-api-access-bllfl") pod "db0e9ee6-03fb-45da-abda-e9af50b92dc7" (UID: "db0e9ee6-03fb-45da-abda-e9af50b92dc7"). InnerVolumeSpecName "kube-api-access-bllfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.161495 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bllfl\" (UniqueName: \"kubernetes.io/projected/db0e9ee6-03fb-45da-abda-e9af50b92dc7-kube-api-access-bllfl\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.454988 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.458374 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.467130 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-combined-ca-bundle\") pod \"e4537dd0-c22a-4879-9fa2-d02f3403172d\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.467250 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-ovndb-tls-certs\") pod \"e4537dd0-c22a-4879-9fa2-d02f3403172d\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.467478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnthx\" (UniqueName: \"kubernetes.io/projected/e4537dd0-c22a-4879-9fa2-d02f3403172d-kube-api-access-lnthx\") pod \"e4537dd0-c22a-4879-9fa2-d02f3403172d\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.467517 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-httpd-config\") pod \"e4537dd0-c22a-4879-9fa2-d02f3403172d\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.467577 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-config\") pod \"e4537dd0-c22a-4879-9fa2-d02f3403172d\" (UID: \"e4537dd0-c22a-4879-9fa2-d02f3403172d\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.481768 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e4537dd0-c22a-4879-9fa2-d02f3403172d" (UID: "e4537dd0-c22a-4879-9fa2-d02f3403172d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.486956 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4537dd0-c22a-4879-9fa2-d02f3403172d-kube-api-access-lnthx" (OuterVolumeSpecName: "kube-api-access-lnthx") pod "e4537dd0-c22a-4879-9fa2-d02f3403172d" (UID: "e4537dd0-c22a-4879-9fa2-d02f3403172d"). InnerVolumeSpecName "kube-api-access-lnthx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.488237 4926 generic.go:334] "Generic (PLEG): container finished" podID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerID="6e8c53be911ace17038a2e6cadd015035402ab7fdd4bc20dffa0eabf1a87b53e" exitCode=0 Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.488267 4926 generic.go:334] "Generic (PLEG): container finished" podID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerID="f6578f217c33ea24cb13af9db1810f8190e7a8bdcfe37c0c1cd2982992eb3f4f" exitCode=143 Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.488328 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3bdab5b7-41be-4087-b654-fd2434a04aaa","Type":"ContainerDied","Data":"6e8c53be911ace17038a2e6cadd015035402ab7fdd4bc20dffa0eabf1a87b53e"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.488357 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3bdab5b7-41be-4087-b654-fd2434a04aaa","Type":"ContainerDied","Data":"f6578f217c33ea24cb13af9db1810f8190e7a8bdcfe37c0c1cd2982992eb3f4f"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.505485 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8xkbk" event={"ID":"5c212508-2303-4aac-86eb-0c5ee339e517","Type":"ContainerDied","Data":"d2a532f9e15546fd38f8ba9aa0e8c6479db0cdfe5d251cbd5d3ecd76e2537e38"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.505595 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a532f9e15546fd38f8ba9aa0e8c6479db0cdfe5d251cbd5d3ecd76e2537e38" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.505615 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8xkbk" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.511569 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"15c1a57a-dd1c-4955-ab28-86f0d09faffb","Type":"ContainerStarted","Data":"e8648823d49dadf55c9d03382f2739927496590ff856fcff5f79430eb4bd915a"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.521387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-86p2f" event={"ID":"db0e9ee6-03fb-45da-abda-e9af50b92dc7","Type":"ContainerDied","Data":"d724aeee9b82ad493b13697cc0d6fc886f555e52529475d3d83ba6b8f6e12f19"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.521763 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d724aeee9b82ad493b13697cc0d6fc886f555e52529475d3d83ba6b8f6e12f19" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.521546 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-86p2f" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.538829 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.539891 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerStarted","Data":"d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.551508 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.551485755 podStartE2EDuration="4.551485755s" podCreationTimestamp="2025-10-07 21:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:55.530790949 +0000 UTC m=+1145.569372099" watchObservedRunningTime="2025-10-07 21:14:55.551485755 +0000 UTC m=+1145.590066905" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.559860 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e4b6a533-2f48-428f-b735-3ceac2e2d7c7","Type":"ContainerStarted","Data":"83fa779013577b23c401508378fa2a3b3312573356b6a31ea3cbf37265190663"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570264 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnthx\" (UniqueName: \"kubernetes.io/projected/e4537dd0-c22a-4879-9fa2-d02f3403172d-kube-api-access-lnthx\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570303 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570858 4926 generic.go:334] "Generic (PLEG): container finished" podID="3ef93fed-2339-4404-adcb-76fcb021c782" containerID="19b4e58f9466c1b8227f113d5aacfb990e8ca33333ceaf1283f31330a9dd1895" exitCode=0 Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570878 4926 generic.go:334] "Generic (PLEG): container finished" podID="3ef93fed-2339-4404-adcb-76fcb021c782" containerID="c97a1128966df6323a30edc576ce4e36b29dd5694c631091b9201d3843aad377" exitCode=2 Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570886 4926 generic.go:334] "Generic (PLEG): container finished" podID="3ef93fed-2339-4404-adcb-76fcb021c782" containerID="682ec989e0923a62b8b1d4ccbe15cbce926d45bc56fc13bf4a977837b79dbec7" exitCode=0 Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570931 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerDied","Data":"19b4e58f9466c1b8227f113d5aacfb990e8ca33333ceaf1283f31330a9dd1895"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570979 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerDied","Data":"c97a1128966df6323a30edc576ce4e36b29dd5694c631091b9201d3843aad377"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.570991 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerDied","Data":"682ec989e0923a62b8b1d4ccbe15cbce926d45bc56fc13bf4a977837b79dbec7"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.572306 4926 generic.go:334] "Generic (PLEG): container finished" podID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerID="f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a" exitCode=0 Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.572358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df5c749bd-sp7b6" event={"ID":"e4537dd0-c22a-4879-9fa2-d02f3403172d","Type":"ContainerDied","Data":"f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.572508 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df5c749bd-sp7b6" event={"ID":"e4537dd0-c22a-4879-9fa2-d02f3403172d","Type":"ContainerDied","Data":"383b5077ce69a5fe963cf6a75b6ac4a4dcf5d2afa9e458f126cd5305e1699950"} Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.572537 4926 scope.go:117] "RemoveContainer" containerID="eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.572695 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df5c749bd-sp7b6" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.641478 4926 scope.go:117] "RemoveContainer" containerID="f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.660443 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e4537dd0-c22a-4879-9fa2-d02f3403172d" (UID: "e4537dd0-c22a-4879-9fa2-d02f3403172d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.661146 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4537dd0-c22a-4879-9fa2-d02f3403172d" (UID: "e4537dd0-c22a-4879-9fa2-d02f3403172d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.670944 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.677903 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-combined-ca-bundle\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.677942 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-scripts\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678002 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq6ph\" (UniqueName: \"kubernetes.io/projected/3bdab5b7-41be-4087-b654-fd2434a04aaa-kube-api-access-zq6ph\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678029 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data-custom\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678075 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bdab5b7-41be-4087-b654-fd2434a04aaa-etc-machine-id\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678125 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bdab5b7-41be-4087-b654-fd2434a04aaa-logs\") pod \"3bdab5b7-41be-4087-b654-fd2434a04aaa\" (UID: \"3bdab5b7-41be-4087-b654-fd2434a04aaa\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678147 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2mjz\" (UniqueName: \"kubernetes.io/projected/5c212508-2303-4aac-86eb-0c5ee339e517-kube-api-access-b2mjz\") pod \"5c212508-2303-4aac-86eb-0c5ee339e517\" (UID: \"5c212508-2303-4aac-86eb-0c5ee339e517\") " Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678444 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bdab5b7-41be-4087-b654-fd2434a04aaa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678577 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bdab5b7-41be-4087-b654-fd2434a04aaa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678641 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.678693 4926 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.681067 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bdab5b7-41be-4087-b654-fd2434a04aaa-logs" (OuterVolumeSpecName: "logs") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.685399 4926 scope.go:117] "RemoveContainer" containerID="eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.685548 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-scripts" (OuterVolumeSpecName: "scripts") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.685904 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c212508-2303-4aac-86eb-0c5ee339e517-kube-api-access-b2mjz" (OuterVolumeSpecName: "kube-api-access-b2mjz") pod "5c212508-2303-4aac-86eb-0c5ee339e517" (UID: "5c212508-2303-4aac-86eb-0c5ee339e517"). InnerVolumeSpecName "kube-api-access-b2mjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: E1007 21:14:55.686021 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31\": container with ID starting with eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31 not found: ID does not exist" containerID="eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.686051 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31"} err="failed to get container status \"eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31\": rpc error: code = NotFound desc = could not find container \"eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31\": container with ID starting with eb4037b12369728731ce5508a82321b893b882cfe136536f1ae2e2e1de44de31 not found: ID does not exist" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.686075 4926 scope.go:117] "RemoveContainer" containerID="f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a" Oct 07 21:14:55 crc kubenswrapper[4926]: E1007 21:14:55.686426 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a\": container with ID starting with f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a not found: ID does not exist" containerID="f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.686746 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a"} err="failed to get container status \"f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a\": rpc error: code = NotFound desc = could not find container \"f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a\": container with ID starting with f0f7b7875c83d03bf1892d2fe95fb6269b9ccbce50ea3eaece25e1b2e0c2ed2a not found: ID does not exist" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.687352 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bdab5b7-41be-4087-b654-fd2434a04aaa-kube-api-access-zq6ph" (OuterVolumeSpecName: "kube-api-access-zq6ph") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "kube-api-access-zq6ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.692284 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.696341 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-config" (OuterVolumeSpecName: "config") pod "e4537dd0-c22a-4879-9fa2-d02f3403172d" (UID: "e4537dd0-c22a-4879-9fa2-d02f3403172d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.720892 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.748558 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data" (OuterVolumeSpecName: "config-data") pod "3bdab5b7-41be-4087-b654-fd2434a04aaa" (UID: "3bdab5b7-41be-4087-b654-fd2434a04aaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781589 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq6ph\" (UniqueName: \"kubernetes.io/projected/3bdab5b7-41be-4087-b654-fd2434a04aaa-kube-api-access-zq6ph\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781631 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781644 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4537dd0-c22a-4879-9fa2-d02f3403172d-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781656 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bdab5b7-41be-4087-b654-fd2434a04aaa-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781667 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2mjz\" (UniqueName: \"kubernetes.io/projected/5c212508-2303-4aac-86eb-0c5ee339e517-kube-api-access-b2mjz\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781675 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781683 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.781691 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bdab5b7-41be-4087-b654-fd2434a04aaa-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.913124 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5df5c749bd-sp7b6"] Oct 07 21:14:55 crc kubenswrapper[4926]: I1007 21:14:55.934634 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5df5c749bd-sp7b6"] Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.586072 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3bdab5b7-41be-4087-b654-fd2434a04aaa","Type":"ContainerDied","Data":"60d9c7c577cc7229bc0d0d987ad45dfc60f47dda6eaf86edb80adec9efeb1faf"} Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.586392 4926 scope.go:117] "RemoveContainer" containerID="6e8c53be911ace17038a2e6cadd015035402ab7fdd4bc20dffa0eabf1a87b53e" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.586560 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.589961 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e4b6a533-2f48-428f-b735-3ceac2e2d7c7","Type":"ContainerStarted","Data":"8c5d235133275fce1722cd18c938457afebd55ab1bf760271f92868191688156"} Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.589999 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e4b6a533-2f48-428f-b735-3ceac2e2d7c7","Type":"ContainerStarted","Data":"9563c3712a0d8ce00570ade44affcf05d9d7a8da9d3584d898b55b8be8656c0b"} Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.621412 4926 scope.go:117] "RemoveContainer" containerID="f6578f217c33ea24cb13af9db1810f8190e7a8bdcfe37c0c1cd2982992eb3f4f" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.621987 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.621961202 podStartE2EDuration="3.621961202s" podCreationTimestamp="2025-10-07 21:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:56.621729085 +0000 UTC m=+1146.660310235" watchObservedRunningTime="2025-10-07 21:14:56.621961202 +0000 UTC m=+1146.660542352" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.644167 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.660695 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.677666 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678551 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c212508-2303-4aac-86eb-0c5ee339e517" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678580 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c212508-2303-4aac-86eb-0c5ee339e517" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678593 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8039b214-d022-4525-9ae6-c7b11921aa82" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678602 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8039b214-d022-4525-9ae6-c7b11921aa82" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678619 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-httpd" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678630 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-httpd" Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678666 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-api" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678673 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-api" Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678693 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0e9ee6-03fb-45da-abda-e9af50b92dc7" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678701 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0e9ee6-03fb-45da-abda-e9af50b92dc7" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678720 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api-log" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678731 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api-log" Oct 07 21:14:56 crc kubenswrapper[4926]: E1007 21:14:56.678764 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.678773 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679007 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0e9ee6-03fb-45da-abda-e9af50b92dc7" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679029 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8039b214-d022-4525-9ae6-c7b11921aa82" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679047 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api-log" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679064 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-api" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679082 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" containerName="cinder-api" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679109 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c212508-2303-4aac-86eb-0c5ee339e517" containerName="mariadb-database-create" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.679127 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" containerName="neutron-httpd" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.681372 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.686555 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.686776 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.686925 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.716936 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bdab5b7-41be-4087-b654-fd2434a04aaa" path="/var/lib/kubelet/pods/3bdab5b7-41be-4087-b654-fd2434a04aaa/volumes" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.718745 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4537dd0-c22a-4879-9fa2-d02f3403172d" path="/var/lib/kubelet/pods/e4537dd0-c22a-4879-9fa2-d02f3403172d/volumes" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.719380 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806182 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806414 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-scripts\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806550 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkj2f\" (UniqueName: \"kubernetes.io/projected/aabe891b-eda8-4359-b8c8-d4d99e309148-kube-api-access-mkj2f\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aabe891b-eda8-4359-b8c8-d4d99e309148-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806678 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabe891b-eda8-4359-b8c8-d4d99e309148-logs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806755 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-public-tls-certs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806797 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-config-data\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.806901 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-config-data-custom\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.807002 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908254 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkj2f\" (UniqueName: \"kubernetes.io/projected/aabe891b-eda8-4359-b8c8-d4d99e309148-kube-api-access-mkj2f\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908300 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aabe891b-eda8-4359-b8c8-d4d99e309148-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908329 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabe891b-eda8-4359-b8c8-d4d99e309148-logs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-public-tls-certs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908477 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-config-data\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908525 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-config-data-custom\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908554 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908452 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aabe891b-eda8-4359-b8c8-d4d99e309148-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.908792 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabe891b-eda8-4359-b8c8-d4d99e309148-logs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.909625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.909712 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-scripts\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.913675 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-config-data-custom\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.914523 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-public-tls-certs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.914910 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.915161 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-config-data\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.916475 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.925562 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkj2f\" (UniqueName: \"kubernetes.io/projected/aabe891b-eda8-4359-b8c8-d4d99e309148-kube-api-access-mkj2f\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:56 crc kubenswrapper[4926]: I1007 21:14:56.926041 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aabe891b-eda8-4359-b8c8-d4d99e309148-scripts\") pod \"cinder-api-0\" (UID: \"aabe891b-eda8-4359-b8c8-d4d99e309148\") " pod="openstack/cinder-api-0" Oct 07 21:14:57 crc kubenswrapper[4926]: I1007 21:14:57.029004 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 21:14:57 crc kubenswrapper[4926]: I1007 21:14:57.393551 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-55b995649c-qwlp9" Oct 07 21:14:57 crc kubenswrapper[4926]: I1007 21:14:57.535506 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 21:14:57 crc kubenswrapper[4926]: W1007 21:14:57.542806 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaabe891b_eda8_4359_b8c8_d4d99e309148.slice/crio-2b19b5d6f8ba2025dc92e919f8c3254178eec2c74503d93d932101e6e41b8490 WatchSource:0}: Error finding container 2b19b5d6f8ba2025dc92e919f8c3254178eec2c74503d93d932101e6e41b8490: Status 404 returned error can't find the container with id 2b19b5d6f8ba2025dc92e919f8c3254178eec2c74503d93d932101e6e41b8490 Oct 07 21:14:57 crc kubenswrapper[4926]: I1007 21:14:57.610569 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aabe891b-eda8-4359-b8c8-d4d99e309148","Type":"ContainerStarted","Data":"2b19b5d6f8ba2025dc92e919f8c3254178eec2c74503d93d932101e6e41b8490"} Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.075963 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.245754 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.251174 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.316568 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7b967f55-42zm8"] Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.316832 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerName="dnsmasq-dns" containerID="cri-o://debb1ed18035e660e47af53e537283270b1b3486e56b09532a96eca19b69100a" gracePeriod=10 Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.638572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aabe891b-eda8-4359-b8c8-d4d99e309148","Type":"ContainerStarted","Data":"b18f00cb54e8fbb7d22622152edd0fe4d0548b5b4de7a0a74b0c7260945f86fd"} Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.661023 4926 generic.go:334] "Generic (PLEG): container finished" podID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerID="debb1ed18035e660e47af53e537283270b1b3486e56b09532a96eca19b69100a" exitCode=0 Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.661303 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" event={"ID":"334d53a8-d4f2-410f-8620-a489bb7d8261","Type":"ContainerDied","Data":"debb1ed18035e660e47af53e537283270b1b3486e56b09532a96eca19b69100a"} Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.716077 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:14:58 crc kubenswrapper[4926]: I1007 21:14:58.836229 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.003515 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-svc\") pod \"334d53a8-d4f2-410f-8620-a489bb7d8261\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.003575 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-nb\") pod \"334d53a8-d4f2-410f-8620-a489bb7d8261\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.003741 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-swift-storage-0\") pod \"334d53a8-d4f2-410f-8620-a489bb7d8261\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.003935 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-sb\") pod \"334d53a8-d4f2-410f-8620-a489bb7d8261\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.004010 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-config\") pod \"334d53a8-d4f2-410f-8620-a489bb7d8261\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.004046 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hflj4\" (UniqueName: \"kubernetes.io/projected/334d53a8-d4f2-410f-8620-a489bb7d8261-kube-api-access-hflj4\") pod \"334d53a8-d4f2-410f-8620-a489bb7d8261\" (UID: \"334d53a8-d4f2-410f-8620-a489bb7d8261\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.009013 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/334d53a8-d4f2-410f-8620-a489bb7d8261-kube-api-access-hflj4" (OuterVolumeSpecName: "kube-api-access-hflj4") pod "334d53a8-d4f2-410f-8620-a489bb7d8261" (UID: "334d53a8-d4f2-410f-8620-a489bb7d8261"). InnerVolumeSpecName "kube-api-access-hflj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.072497 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "334d53a8-d4f2-410f-8620-a489bb7d8261" (UID: "334d53a8-d4f2-410f-8620-a489bb7d8261"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.076439 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "334d53a8-d4f2-410f-8620-a489bb7d8261" (UID: "334d53a8-d4f2-410f-8620-a489bb7d8261"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.086396 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "334d53a8-d4f2-410f-8620-a489bb7d8261" (UID: "334d53a8-d4f2-410f-8620-a489bb7d8261"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.089512 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "334d53a8-d4f2-410f-8620-a489bb7d8261" (UID: "334d53a8-d4f2-410f-8620-a489bb7d8261"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.100479 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-config" (OuterVolumeSpecName: "config") pod "334d53a8-d4f2-410f-8620-a489bb7d8261" (UID: "334d53a8-d4f2-410f-8620-a489bb7d8261"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.106837 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.106874 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hflj4\" (UniqueName: \"kubernetes.io/projected/334d53a8-d4f2-410f-8620-a489bb7d8261-kube-api-access-hflj4\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.106886 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.106896 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.106907 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.106916 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/334d53a8-d4f2-410f-8620-a489bb7d8261-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.259100 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7d27-account-create-8482b"] Oct 07 21:14:59 crc kubenswrapper[4926]: E1007 21:14:59.259546 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerName="dnsmasq-dns" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.259565 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerName="dnsmasq-dns" Oct 07 21:14:59 crc kubenswrapper[4926]: E1007 21:14:59.259587 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerName="init" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.259597 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerName="init" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.259813 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" containerName="dnsmasq-dns" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.260658 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.262854 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.269131 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d27-account-create-8482b"] Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.414074 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95hlm\" (UniqueName: \"kubernetes.io/projected/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7-kube-api-access-95hlm\") pod \"nova-api-7d27-account-create-8482b\" (UID: \"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7\") " pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.486914 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-1475-account-create-bxwm8"] Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.489029 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.492815 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.498021 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1475-account-create-bxwm8"] Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.515986 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95hlm\" (UniqueName: \"kubernetes.io/projected/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7-kube-api-access-95hlm\") pod \"nova-api-7d27-account-create-8482b\" (UID: \"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7\") " pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.516067 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8zhx\" (UniqueName: \"kubernetes.io/projected/13c7f3b3-0df4-49d5-a538-0fb9a5e10186-kube-api-access-h8zhx\") pod \"nova-cell0-1475-account-create-bxwm8\" (UID: \"13c7f3b3-0df4-49d5-a538-0fb9a5e10186\") " pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.548917 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95hlm\" (UniqueName: \"kubernetes.io/projected/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7-kube-api-access-95hlm\") pod \"nova-api-7d27-account-create-8482b\" (UID: \"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7\") " pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.618122 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8zhx\" (UniqueName: \"kubernetes.io/projected/13c7f3b3-0df4-49d5-a538-0fb9a5e10186-kube-api-access-h8zhx\") pod \"nova-cell0-1475-account-create-bxwm8\" (UID: \"13c7f3b3-0df4-49d5-a538-0fb9a5e10186\") " pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.619268 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.636832 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8zhx\" (UniqueName: \"kubernetes.io/projected/13c7f3b3-0df4-49d5-a538-0fb9a5e10186-kube-api-access-h8zhx\") pod \"nova-cell0-1475-account-create-bxwm8\" (UID: \"13c7f3b3-0df4-49d5-a538-0fb9a5e10186\") " pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.690229 4926 generic.go:334] "Generic (PLEG): container finished" podID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" exitCode=1 Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.690326 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerDied","Data":"d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb"} Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.690374 4926 scope.go:117] "RemoveContainer" containerID="08b750180d7a5c997bcd1aba25d67cffc8dbf0cb8e825e537fad748c061cc8d9" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.691174 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:14:59 crc kubenswrapper[4926]: E1007 21:14:59.691598 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.704854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aabe891b-eda8-4359-b8c8-d4d99e309148","Type":"ContainerStarted","Data":"5e583032450ce42f0afca6e873d5f1656bedd542159d8188283817a0e219d1d2"} Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.705228 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.713267 4926 generic.go:334] "Generic (PLEG): container finished" podID="3ef93fed-2339-4404-adcb-76fcb021c782" containerID="c2978894f94e376ff151ea0b1c8a3968b11fa5960e7311b617c92b908f8e8925" exitCode=0 Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.713340 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerDied","Data":"c2978894f94e376ff151ea0b1c8a3968b11fa5960e7311b617c92b908f8e8925"} Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.713372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ef93fed-2339-4404-adcb-76fcb021c782","Type":"ContainerDied","Data":"a96ceed949664e3e257628a0dcba5dfd61b3fa7ab6ac5efb8b137cc36f4254ab"} Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.713382 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a96ceed949664e3e257628a0dcba5dfd61b3fa7ab6ac5efb8b137cc36f4254ab" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.719036 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="cinder-scheduler" containerID="cri-o://b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71" gracePeriod=30 Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.719162 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.719526 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7b967f55-42zm8" event={"ID":"334d53a8-d4f2-410f-8620-a489bb7d8261","Type":"ContainerDied","Data":"365f0c6abc27888799fc58a3ff9cf1a39c7d61c1d4441728af389fb6b870556f"} Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.719650 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="probe" containerID="cri-o://ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a" gracePeriod=30 Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.783425 4926 scope.go:117] "RemoveContainer" containerID="debb1ed18035e660e47af53e537283270b1b3486e56b09532a96eca19b69100a" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.788605 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.792313 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.792286253 podStartE2EDuration="3.792286253s" podCreationTimestamp="2025-10-07 21:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:14:59.748719027 +0000 UTC m=+1149.787300177" watchObservedRunningTime="2025-10-07 21:14:59.792286253 +0000 UTC m=+1149.830867403" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.797441 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7b967f55-42zm8"] Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.821807 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7b967f55-42zm8"] Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.822301 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831019 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-log-httpd\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831115 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-sg-core-conf-yaml\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831161 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-config-data\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831176 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-combined-ca-bundle\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831221 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-scripts\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831265 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp6zh\" (UniqueName: \"kubernetes.io/projected/3ef93fed-2339-4404-adcb-76fcb021c782-kube-api-access-cp6zh\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831289 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-run-httpd\") pod \"3ef93fed-2339-4404-adcb-76fcb021c782\" (UID: \"3ef93fed-2339-4404-adcb-76fcb021c782\") " Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.831778 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.832089 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.833429 4926 scope.go:117] "RemoveContainer" containerID="5b31e20cf3b6d4e484a1e2595386f0f383cb4de11a5c468db2158bd10cb58258" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.837210 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef93fed-2339-4404-adcb-76fcb021c782-kube-api-access-cp6zh" (OuterVolumeSpecName: "kube-api-access-cp6zh") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "kube-api-access-cp6zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.837825 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-scripts" (OuterVolumeSpecName: "scripts") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.907073 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.934295 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.934343 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.934356 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.934364 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp6zh\" (UniqueName: \"kubernetes.io/projected/3ef93fed-2339-4404-adcb-76fcb021c782-kube-api-access-cp6zh\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.934374 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ef93fed-2339-4404-adcb-76fcb021c782-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:14:59 crc kubenswrapper[4926]: I1007 21:14:59.982834 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.017038 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-config-data" (OuterVolumeSpecName: "config-data") pod "3ef93fed-2339-4404-adcb-76fcb021c782" (UID: "3ef93fed-2339-4404-adcb-76fcb021c782"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.037574 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.037605 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef93fed-2339-4404-adcb-76fcb021c782-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.151284 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l"] Oct 07 21:15:00 crc kubenswrapper[4926]: E1007 21:15:00.151840 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="sg-core" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.151859 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="sg-core" Oct 07 21:15:00 crc kubenswrapper[4926]: E1007 21:15:00.151868 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-central-agent" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.151875 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-central-agent" Oct 07 21:15:00 crc kubenswrapper[4926]: E1007 21:15:00.151901 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="proxy-httpd" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.151908 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="proxy-httpd" Oct 07 21:15:00 crc kubenswrapper[4926]: E1007 21:15:00.151928 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-notification-agent" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.151935 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-notification-agent" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.152129 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-central-agent" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.152143 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="proxy-httpd" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.152150 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="ceilometer-notification-agent" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.152174 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" containerName="sg-core" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.152928 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.155757 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.156019 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.167825 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l"] Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.206674 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d27-account-create-8482b"] Oct 07 21:15:00 crc kubenswrapper[4926]: W1007 21:15:00.210661 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3890b5d3_dbd2_4c4f_86c1_7c52a37a65f7.slice/crio-bc77b0ec95c450329d5f700c13bcf0ebfd1b21b83d5bb3ecf236e4d5e75c22d8 WatchSource:0}: Error finding container bc77b0ec95c450329d5f700c13bcf0ebfd1b21b83d5bb3ecf236e4d5e75c22d8: Status 404 returned error can't find the container with id bc77b0ec95c450329d5f700c13bcf0ebfd1b21b83d5bb3ecf236e4d5e75c22d8 Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.244359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr5m7\" (UniqueName: \"kubernetes.io/projected/1d1e6970-d33f-4582-9485-2a9027cbb1d9-kube-api-access-vr5m7\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.245131 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d1e6970-d33f-4582-9485-2a9027cbb1d9-config-volume\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.245264 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d1e6970-d33f-4582-9485-2a9027cbb1d9-secret-volume\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.346772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr5m7\" (UniqueName: \"kubernetes.io/projected/1d1e6970-d33f-4582-9485-2a9027cbb1d9-kube-api-access-vr5m7\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.346851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d1e6970-d33f-4582-9485-2a9027cbb1d9-config-volume\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.346871 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d1e6970-d33f-4582-9485-2a9027cbb1d9-secret-volume\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.351930 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d1e6970-d33f-4582-9485-2a9027cbb1d9-config-volume\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.355415 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d1e6970-d33f-4582-9485-2a9027cbb1d9-secret-volume\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.372144 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr5m7\" (UniqueName: \"kubernetes.io/projected/1d1e6970-d33f-4582-9485-2a9027cbb1d9-kube-api-access-vr5m7\") pod \"collect-profiles-29331195-vw22l\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.418415 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1475-account-create-bxwm8"] Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.488917 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.702384 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="334d53a8-d4f2-410f-8620-a489bb7d8261" path="/var/lib/kubelet/pods/334d53a8-d4f2-410f-8620-a489bb7d8261/volumes" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.762212 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1475-account-create-bxwm8" event={"ID":"13c7f3b3-0df4-49d5-a538-0fb9a5e10186","Type":"ContainerStarted","Data":"f8a15671a240c2ba3ea9866ccc027966bed8fceb49d9bb321f03738243ec4eb3"} Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.765499 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d27-account-create-8482b" event={"ID":"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7","Type":"ContainerStarted","Data":"b3d64569500e4791b31a167e08b98cf3b3edff5025edc1d41ca623a28eee7653"} Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.765556 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d27-account-create-8482b" event={"ID":"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7","Type":"ContainerStarted","Data":"bc77b0ec95c450329d5f700c13bcf0ebfd1b21b83d5bb3ecf236e4d5e75c22d8"} Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.772757 4926 generic.go:334] "Generic (PLEG): container finished" podID="dca9ab35-65a2-4034-921a-a295cde7e581" containerID="ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a" exitCode=0 Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.773883 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dca9ab35-65a2-4034-921a-a295cde7e581","Type":"ContainerDied","Data":"ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a"} Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.774041 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.837280 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.847654 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.856682 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.859243 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.862664 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.862842 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.869297 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.947282 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.947319 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.947334 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.947346 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:00 crc kubenswrapper[4926]: I1007 21:15:00.948031 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:15:00 crc kubenswrapper[4926]: E1007 21:15:00.948294 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.054967 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l"] Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.073831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-run-httpd\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.073871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.073894 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-scripts\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.074128 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv7n5\" (UniqueName: \"kubernetes.io/projected/24fe47d6-775e-43a2-b488-3bdd9518d4a8-kube-api-access-rv7n5\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.074197 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.074298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-config-data\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.074331 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-log-httpd\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.176618 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-run-httpd\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.176760 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.176834 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-scripts\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.176935 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv7n5\" (UniqueName: \"kubernetes.io/projected/24fe47d6-775e-43a2-b488-3bdd9518d4a8-kube-api-access-rv7n5\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.177033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.177118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-config-data\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.177184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-log-httpd\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.177666 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-log-httpd\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.177921 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-run-httpd\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.185077 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.185347 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-config-data\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.185599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-scripts\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.185709 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.202746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv7n5\" (UniqueName: \"kubernetes.io/projected/24fe47d6-775e-43a2-b488-3bdd9518d4a8-kube-api-access-rv7n5\") pod \"ceilometer-0\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.484157 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.601606 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.601655 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 21:15:01 crc kubenswrapper[4926]: E1007 21:15:01.632854 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47c8f35_5fce_4dc5_b86d_895606525d5f.slice/crio-37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d1e6970_d33f_4582_9485_2a9027cbb1d9.slice/crio-3a3fd12ee6604659fb4eaa418630e6d708a54bd07d2a53fe165f8d96fe374542.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d1e6970_d33f_4582_9485_2a9027cbb1d9.slice/crio-conmon-3a3fd12ee6604659fb4eaa418630e6d708a54bd07d2a53fe165f8d96fe374542.scope\": RecentStats: unable to find data in memory cache]" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.660981 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.674985 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.718282 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.794232 4926 generic.go:334] "Generic (PLEG): container finished" podID="dca9ab35-65a2-4034-921a-a295cde7e581" containerID="b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71" exitCode=0 Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.794306 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dca9ab35-65a2-4034-921a-a295cde7e581","Type":"ContainerDied","Data":"b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71"} Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.794342 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dca9ab35-65a2-4034-921a-a295cde7e581","Type":"ContainerDied","Data":"ccb840e6ebdc48bcdc07f9bb76847a6138f6773c08de846ae78279ce55616146"} Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.794358 4926 scope.go:117] "RemoveContainer" containerID="ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.794513 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.807874 4926 generic.go:334] "Generic (PLEG): container finished" podID="3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7" containerID="b3d64569500e4791b31a167e08b98cf3b3edff5025edc1d41ca623a28eee7653" exitCode=0 Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.807973 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d27-account-create-8482b" event={"ID":"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7","Type":"ContainerDied","Data":"b3d64569500e4791b31a167e08b98cf3b3edff5025edc1d41ca623a28eee7653"} Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.813454 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d1e6970-d33f-4582-9485-2a9027cbb1d9" containerID="3a3fd12ee6604659fb4eaa418630e6d708a54bd07d2a53fe165f8d96fe374542" exitCode=0 Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.813567 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" event={"ID":"1d1e6970-d33f-4582-9485-2a9027cbb1d9","Type":"ContainerDied","Data":"3a3fd12ee6604659fb4eaa418630e6d708a54bd07d2a53fe165f8d96fe374542"} Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.813598 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" event={"ID":"1d1e6970-d33f-4582-9485-2a9027cbb1d9","Type":"ContainerStarted","Data":"da7b997f75d0c5de72f79389c1934b64e6c8eaad15acd2ef82e61fe30960303c"} Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.817857 4926 generic.go:334] "Generic (PLEG): container finished" podID="13c7f3b3-0df4-49d5-a538-0fb9a5e10186" containerID="66d40c06f81fabf98b39eebfd6c0b56ba56d78593a1d36407ecf365eb6f9501a" exitCode=0 Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.818467 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1475-account-create-bxwm8" event={"ID":"13c7f3b3-0df4-49d5-a538-0fb9a5e10186","Type":"ContainerDied","Data":"66d40c06f81fabf98b39eebfd6c0b56ba56d78593a1d36407ecf365eb6f9501a"} Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.818559 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.818931 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.841648 4926 scope.go:117] "RemoveContainer" containerID="b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.867333 4926 scope.go:117] "RemoveContainer" containerID="ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a" Oct 07 21:15:01 crc kubenswrapper[4926]: E1007 21:15:01.867930 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a\": container with ID starting with ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a not found: ID does not exist" containerID="ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.867960 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a"} err="failed to get container status \"ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a\": rpc error: code = NotFound desc = could not find container \"ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a\": container with ID starting with ae11e1f83398c75442f5dc75e95943b79bc1d5c6a4cc8b46e8aba4185911939a not found: ID does not exist" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.867985 4926 scope.go:117] "RemoveContainer" containerID="b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71" Oct 07 21:15:01 crc kubenswrapper[4926]: E1007 21:15:01.868415 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71\": container with ID starting with b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71 not found: ID does not exist" containerID="b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.868436 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71"} err="failed to get container status \"b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71\": rpc error: code = NotFound desc = could not find container \"b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71\": container with ID starting with b46b421211c4a37191db262e2e9745b47faafc58a6f03f3ea998db5b882fab71 not found: ID does not exist" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.902557 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntkgd\" (UniqueName: \"kubernetes.io/projected/dca9ab35-65a2-4034-921a-a295cde7e581-kube-api-access-ntkgd\") pod \"dca9ab35-65a2-4034-921a-a295cde7e581\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.902863 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data\") pod \"dca9ab35-65a2-4034-921a-a295cde7e581\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.903029 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data-custom\") pod \"dca9ab35-65a2-4034-921a-a295cde7e581\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.903064 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-combined-ca-bundle\") pod \"dca9ab35-65a2-4034-921a-a295cde7e581\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.903154 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-scripts\") pod \"dca9ab35-65a2-4034-921a-a295cde7e581\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.903215 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dca9ab35-65a2-4034-921a-a295cde7e581-etc-machine-id\") pod \"dca9ab35-65a2-4034-921a-a295cde7e581\" (UID: \"dca9ab35-65a2-4034-921a-a295cde7e581\") " Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.903752 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dca9ab35-65a2-4034-921a-a295cde7e581-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dca9ab35-65a2-4034-921a-a295cde7e581" (UID: "dca9ab35-65a2-4034-921a-a295cde7e581"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.904711 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dca9ab35-65a2-4034-921a-a295cde7e581-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.910531 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-scripts" (OuterVolumeSpecName: "scripts") pod "dca9ab35-65a2-4034-921a-a295cde7e581" (UID: "dca9ab35-65a2-4034-921a-a295cde7e581"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.910634 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dca9ab35-65a2-4034-921a-a295cde7e581-kube-api-access-ntkgd" (OuterVolumeSpecName: "kube-api-access-ntkgd") pod "dca9ab35-65a2-4034-921a-a295cde7e581" (UID: "dca9ab35-65a2-4034-921a-a295cde7e581"). InnerVolumeSpecName "kube-api-access-ntkgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:01 crc kubenswrapper[4926]: I1007 21:15:01.960415 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dca9ab35-65a2-4034-921a-a295cde7e581" (UID: "dca9ab35-65a2-4034-921a-a295cde7e581"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.011670 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntkgd\" (UniqueName: \"kubernetes.io/projected/dca9ab35-65a2-4034-921a-a295cde7e581-kube-api-access-ntkgd\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.012150 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.012162 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.016610 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.035395 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dca9ab35-65a2-4034-921a-a295cde7e581" (UID: "dca9ab35-65a2-4034-921a-a295cde7e581"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.054466 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data" (OuterVolumeSpecName: "config-data") pod "dca9ab35-65a2-4034-921a-a295cde7e581" (UID: "dca9ab35-65a2-4034-921a-a295cde7e581"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.113810 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.113839 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca9ab35-65a2-4034-921a-a295cde7e581-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.149166 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.162288 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.165573 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.176447 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:15:02 crc kubenswrapper[4926]: E1007 21:15:02.177049 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="probe" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.177062 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="probe" Oct 07 21:15:02 crc kubenswrapper[4926]: E1007 21:15:02.177087 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="cinder-scheduler" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.177094 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="cinder-scheduler" Oct 07 21:15:02 crc kubenswrapper[4926]: E1007 21:15:02.177120 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7" containerName="mariadb-account-create" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.177126 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7" containerName="mariadb-account-create" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.177398 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="cinder-scheduler" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.177425 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" containerName="probe" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.177443 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7" containerName="mariadb-account-create" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.178570 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.181726 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.189244 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.317259 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95hlm\" (UniqueName: \"kubernetes.io/projected/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7-kube-api-access-95hlm\") pod \"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7\" (UID: \"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7\") " Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.317590 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-config-data\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.317630 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbfp6\" (UniqueName: \"kubernetes.io/projected/348d1b67-e2d4-4699-9890-3e3befa0133c-kube-api-access-mbfp6\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.317654 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.317701 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-scripts\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.318019 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/348d1b67-e2d4-4699-9890-3e3befa0133c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.318090 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.320812 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7-kube-api-access-95hlm" (OuterVolumeSpecName: "kube-api-access-95hlm") pod "3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7" (UID: "3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7"). InnerVolumeSpecName "kube-api-access-95hlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.419872 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-config-data\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.419916 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbfp6\" (UniqueName: \"kubernetes.io/projected/348d1b67-e2d4-4699-9890-3e3befa0133c-kube-api-access-mbfp6\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.419939 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.419976 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-scripts\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.420040 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/348d1b67-e2d4-4699-9890-3e3befa0133c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.420065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.420106 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/348d1b67-e2d4-4699-9890-3e3befa0133c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.420181 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95hlm\" (UniqueName: \"kubernetes.io/projected/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7-kube-api-access-95hlm\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.431186 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-config-data\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.432195 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.432310 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-scripts\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.437716 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/348d1b67-e2d4-4699-9890-3e3befa0133c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.437737 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbfp6\" (UniqueName: \"kubernetes.io/projected/348d1b67-e2d4-4699-9890-3e3befa0133c-kube-api-access-mbfp6\") pod \"cinder-scheduler-0\" (UID: \"348d1b67-e2d4-4699-9890-3e3befa0133c\") " pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.520267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.695173 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ef93fed-2339-4404-adcb-76fcb021c782" path="/var/lib/kubelet/pods/3ef93fed-2339-4404-adcb-76fcb021c782/volumes" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.696306 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dca9ab35-65a2-4034-921a-a295cde7e581" path="/var/lib/kubelet/pods/dca9ab35-65a2-4034-921a-a295cde7e581/volumes" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.837476 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerStarted","Data":"9dfad846bf8620ec2edc170dc1fb8fef4c604838d0b8e64c2f6751367310fe12"} Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.837890 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerStarted","Data":"78b4437ff7fba882a6551eddf2b75165e816cecad0c99a8d80e87bf71d03c100"} Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.837909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerStarted","Data":"df06be170555421c7c21b2f392545eb9f1051ba475e563ccf15dcd09d1d909b7"} Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.838829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d27-account-create-8482b" event={"ID":"3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7","Type":"ContainerDied","Data":"bc77b0ec95c450329d5f700c13bcf0ebfd1b21b83d5bb3ecf236e4d5e75c22d8"} Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.838865 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc77b0ec95c450329d5f700c13bcf0ebfd1b21b83d5bb3ecf236e4d5e75c22d8" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.838863 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d27-account-create-8482b" Oct 07 21:15:02 crc kubenswrapper[4926]: I1007 21:15:02.987874 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.213700 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.214003 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.214074 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.214976 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c2c0ad449595a88cf92d0a873021ac44f258f8c38778755a129cc3c50f40b2f"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.215045 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://4c2c0ad449595a88cf92d0a873021ac44f258f8c38778755a129cc3c50f40b2f" gracePeriod=600 Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.378706 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.548074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8zhx\" (UniqueName: \"kubernetes.io/projected/13c7f3b3-0df4-49d5-a538-0fb9a5e10186-kube-api-access-h8zhx\") pod \"13c7f3b3-0df4-49d5-a538-0fb9a5e10186\" (UID: \"13c7f3b3-0df4-49d5-a538-0fb9a5e10186\") " Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.557773 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c7f3b3-0df4-49d5-a538-0fb9a5e10186-kube-api-access-h8zhx" (OuterVolumeSpecName: "kube-api-access-h8zhx") pod "13c7f3b3-0df4-49d5-a538-0fb9a5e10186" (UID: "13c7f3b3-0df4-49d5-a538-0fb9a5e10186"). InnerVolumeSpecName "kube-api-access-h8zhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.650450 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8zhx\" (UniqueName: \"kubernetes.io/projected/13c7f3b3-0df4-49d5-a538-0fb9a5e10186-kube-api-access-h8zhx\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.713247 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.775923 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.775975 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.833600 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.856612 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d1e6970-d33f-4582-9485-2a9027cbb1d9-config-volume\") pod \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.856688 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr5m7\" (UniqueName: \"kubernetes.io/projected/1d1e6970-d33f-4582-9485-2a9027cbb1d9-kube-api-access-vr5m7\") pod \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.856935 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d1e6970-d33f-4582-9485-2a9027cbb1d9-secret-volume\") pod \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\" (UID: \"1d1e6970-d33f-4582-9485-2a9027cbb1d9\") " Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.857638 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d1e6970-d33f-4582-9485-2a9027cbb1d9-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d1e6970-d33f-4582-9485-2a9027cbb1d9" (UID: "1d1e6970-d33f-4582-9485-2a9027cbb1d9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.857918 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d1e6970-d33f-4582-9485-2a9027cbb1d9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.879961 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" event={"ID":"1d1e6970-d33f-4582-9485-2a9027cbb1d9","Type":"ContainerDied","Data":"da7b997f75d0c5de72f79389c1934b64e6c8eaad15acd2ef82e61fe30960303c"} Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.880001 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da7b997f75d0c5de72f79389c1934b64e6c8eaad15acd2ef82e61fe30960303c" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.880057 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.885338 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1475-account-create-bxwm8" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.885667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1475-account-create-bxwm8" event={"ID":"13c7f3b3-0df4-49d5-a538-0fb9a5e10186","Type":"ContainerDied","Data":"f8a15671a240c2ba3ea9866ccc027966bed8fceb49d9bb321f03738243ec4eb3"} Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.886143 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8a15671a240c2ba3ea9866ccc027966bed8fceb49d9bb321f03738243ec4eb3" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.889775 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"348d1b67-e2d4-4699-9890-3e3befa0133c","Type":"ContainerStarted","Data":"1345905be07271b3854f65be87c2241bc27d3abd830e54834b6d3ad16f74ea8a"} Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.889777 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d1e6970-d33f-4582-9485-2a9027cbb1d9-kube-api-access-vr5m7" (OuterVolumeSpecName: "kube-api-access-vr5m7") pod "1d1e6970-d33f-4582-9485-2a9027cbb1d9" (UID: "1d1e6970-d33f-4582-9485-2a9027cbb1d9"). InnerVolumeSpecName "kube-api-access-vr5m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.897972 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="4c2c0ad449595a88cf92d0a873021ac44f258f8c38778755a129cc3c50f40b2f" exitCode=0 Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.898034 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"4c2c0ad449595a88cf92d0a873021ac44f258f8c38778755a129cc3c50f40b2f"} Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.898073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"65db47897203feb0e606b82feeba2549b62a2540d28ad5a38331860f14f09627"} Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.898092 4926 scope.go:117] "RemoveContainer" containerID="a7195a5bc10effcad121b73904803864abd9373d3dda56955bffbf5b0e3d1427" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.899037 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.902096 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1e6970-d33f-4582-9485-2a9027cbb1d9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d1e6970-d33f-4582-9485-2a9027cbb1d9" (UID: "1d1e6970-d33f-4582-9485-2a9027cbb1d9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.909145 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.960987 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr5m7\" (UniqueName: \"kubernetes.io/projected/1d1e6970-d33f-4582-9485-2a9027cbb1d9-kube-api-access-vr5m7\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:03 crc kubenswrapper[4926]: I1007 21:15:03.961019 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d1e6970-d33f-4582-9485-2a9027cbb1d9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.350675 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.350779 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.358055 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.723846 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k99cm"] Oct 07 21:15:04 crc kubenswrapper[4926]: E1007 21:15:04.724821 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c7f3b3-0df4-49d5-a538-0fb9a5e10186" containerName="mariadb-account-create" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.724840 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c7f3b3-0df4-49d5-a538-0fb9a5e10186" containerName="mariadb-account-create" Oct 07 21:15:04 crc kubenswrapper[4926]: E1007 21:15:04.724867 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d1e6970-d33f-4582-9485-2a9027cbb1d9" containerName="collect-profiles" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.724875 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d1e6970-d33f-4582-9485-2a9027cbb1d9" containerName="collect-profiles" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.725063 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c7f3b3-0df4-49d5-a538-0fb9a5e10186" containerName="mariadb-account-create" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.725087 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d1e6970-d33f-4582-9485-2a9027cbb1d9" containerName="collect-profiles" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.725798 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.730471 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5qfg2" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.730663 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.730984 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.753978 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k99cm"] Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.895284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-scripts\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.896394 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.896497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcvgb\" (UniqueName: \"kubernetes.io/projected/7a4ae7fe-c26e-4dca-a071-c03a34582e53-kube-api-access-wcvgb\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.896617 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-config-data\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.940940 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"348d1b67-e2d4-4699-9890-3e3befa0133c","Type":"ContainerStarted","Data":"de2b636d94ea888c7e4a94870200aabf7f342a5c1c7783288cde421f0af38b00"} Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.945265 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerStarted","Data":"3d6ecf23f0e262653ff5e88c263648a6c8a736c25ee03c6ee904b50f2f807d83"} Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.946030 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.998847 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.998890 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcvgb\" (UniqueName: \"kubernetes.io/projected/7a4ae7fe-c26e-4dca-a071-c03a34582e53-kube-api-access-wcvgb\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.998940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-config-data\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:04 crc kubenswrapper[4926]: I1007 21:15:04.998998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-scripts\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.005459 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-config-data\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.007703 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.017767 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcvgb\" (UniqueName: \"kubernetes.io/projected/7a4ae7fe-c26e-4dca-a071-c03a34582e53-kube-api-access-wcvgb\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.023724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-scripts\") pod \"nova-cell0-conductor-db-sync-k99cm\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.055850 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.613851 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k99cm"] Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.975964 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k99cm" event={"ID":"7a4ae7fe-c26e-4dca-a071-c03a34582e53","Type":"ContainerStarted","Data":"c60cb4c34c03fea90a81a1a96b166082a0dc1eb9d680a8e7218ae80c3733fac4"} Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.982485 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 21:15:05 crc kubenswrapper[4926]: I1007 21:15:05.983883 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"348d1b67-e2d4-4699-9890-3e3befa0133c","Type":"ContainerStarted","Data":"fa7850fb1d5e529e528603da4232d16a62a764d58a8a87aa20a1b4b2d14eecc5"} Oct 07 21:15:06 crc kubenswrapper[4926]: I1007 21:15:06.004724 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.004705748 podStartE2EDuration="4.004705748s" podCreationTimestamp="2025-10-07 21:15:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:15:06.003545725 +0000 UTC m=+1156.042126875" watchObservedRunningTime="2025-10-07 21:15:06.004705748 +0000 UTC m=+1156.043286898" Oct 07 21:15:06 crc kubenswrapper[4926]: I1007 21:15:06.451373 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:06 crc kubenswrapper[4926]: I1007 21:15:06.451426 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 21:15:07 crc kubenswrapper[4926]: I1007 21:15:07.005845 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerStarted","Data":"69a7fd2489d265d6a4c5d45d64be0255ddb6c4caecf44732012d6b4bf1ead8a2"} Oct 07 21:15:07 crc kubenswrapper[4926]: I1007 21:15:07.006887 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:15:07 crc kubenswrapper[4926]: I1007 21:15:07.051937 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.358200305 podStartE2EDuration="7.051915913s" podCreationTimestamp="2025-10-07 21:15:00 +0000 UTC" firstStartedPulling="2025-10-07 21:15:02.020668806 +0000 UTC m=+1152.059249956" lastFinishedPulling="2025-10-07 21:15:05.714384414 +0000 UTC m=+1155.752965564" observedRunningTime="2025-10-07 21:15:07.050626186 +0000 UTC m=+1157.089207356" watchObservedRunningTime="2025-10-07 21:15:07.051915913 +0000 UTC m=+1157.090497063" Oct 07 21:15:07 crc kubenswrapper[4926]: I1007 21:15:07.521274 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.594601 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-587b-account-create-csjpm"] Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.597356 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.600309 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.604125 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-587b-account-create-csjpm"] Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.716572 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsdw4\" (UniqueName: \"kubernetes.io/projected/7375476d-b601-4885-8620-4f07b35edd55-kube-api-access-fsdw4\") pod \"nova-cell1-587b-account-create-csjpm\" (UID: \"7375476d-b601-4885-8620-4f07b35edd55\") " pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.734817 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.819889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsdw4\" (UniqueName: \"kubernetes.io/projected/7375476d-b601-4885-8620-4f07b35edd55-kube-api-access-fsdw4\") pod \"nova-cell1-587b-account-create-csjpm\" (UID: \"7375476d-b601-4885-8620-4f07b35edd55\") " pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.843639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsdw4\" (UniqueName: \"kubernetes.io/projected/7375476d-b601-4885-8620-4f07b35edd55-kube-api-access-fsdw4\") pod \"nova-cell1-587b-account-create-csjpm\" (UID: \"7375476d-b601-4885-8620-4f07b35edd55\") " pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:09 crc kubenswrapper[4926]: I1007 21:15:09.928174 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:10 crc kubenswrapper[4926]: I1007 21:15:10.438450 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-587b-account-create-csjpm"] Oct 07 21:15:11 crc kubenswrapper[4926]: I1007 21:15:11.053485 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-587b-account-create-csjpm" event={"ID":"7375476d-b601-4885-8620-4f07b35edd55","Type":"ContainerStarted","Data":"6d411557d5ac86cfe63f36f6598e94665f069d0aa9043a8230d89b17eeb68b85"} Oct 07 21:15:11 crc kubenswrapper[4926]: I1007 21:15:11.568375 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:11 crc kubenswrapper[4926]: I1007 21:15:11.570270 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="proxy-httpd" containerID="cri-o://69a7fd2489d265d6a4c5d45d64be0255ddb6c4caecf44732012d6b4bf1ead8a2" gracePeriod=30 Oct 07 21:15:11 crc kubenswrapper[4926]: I1007 21:15:11.570277 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="sg-core" containerID="cri-o://3d6ecf23f0e262653ff5e88c263648a6c8a736c25ee03c6ee904b50f2f807d83" gracePeriod=30 Oct 07 21:15:11 crc kubenswrapper[4926]: I1007 21:15:11.570613 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-central-agent" containerID="cri-o://78b4437ff7fba882a6551eddf2b75165e816cecad0c99a8d80e87bf71d03c100" gracePeriod=30 Oct 07 21:15:11 crc kubenswrapper[4926]: I1007 21:15:11.570310 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-notification-agent" containerID="cri-o://9dfad846bf8620ec2edc170dc1fb8fef4c604838d0b8e64c2f6751367310fe12" gracePeriod=30 Oct 07 21:15:11 crc kubenswrapper[4926]: E1007 21:15:11.938080 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24fe47d6_775e_43a2_b488_3bdd9518d4a8.slice/crio-conmon-3d6ecf23f0e262653ff5e88c263648a6c8a736c25ee03c6ee904b50f2f807d83.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24fe47d6_775e_43a2_b488_3bdd9518d4a8.slice/crio-conmon-69a7fd2489d265d6a4c5d45d64be0255ddb6c4caecf44732012d6b4bf1ead8a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47c8f35_5fce_4dc5_b86d_895606525d5f.slice/crio-37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69\": RecentStats: unable to find data in memory cache]" Oct 07 21:15:12 crc kubenswrapper[4926]: I1007 21:15:12.069567 4926 generic.go:334] "Generic (PLEG): container finished" podID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerID="69a7fd2489d265d6a4c5d45d64be0255ddb6c4caecf44732012d6b4bf1ead8a2" exitCode=0 Oct 07 21:15:12 crc kubenswrapper[4926]: I1007 21:15:12.069615 4926 generic.go:334] "Generic (PLEG): container finished" podID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerID="3d6ecf23f0e262653ff5e88c263648a6c8a736c25ee03c6ee904b50f2f807d83" exitCode=2 Oct 07 21:15:12 crc kubenswrapper[4926]: I1007 21:15:12.069643 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerDied","Data":"69a7fd2489d265d6a4c5d45d64be0255ddb6c4caecf44732012d6b4bf1ead8a2"} Oct 07 21:15:12 crc kubenswrapper[4926]: I1007 21:15:12.069676 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerDied","Data":"3d6ecf23f0e262653ff5e88c263648a6c8a736c25ee03c6ee904b50f2f807d83"} Oct 07 21:15:12 crc kubenswrapper[4926]: I1007 21:15:12.707839 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 21:15:13 crc kubenswrapper[4926]: I1007 21:15:13.082946 4926 generic.go:334] "Generic (PLEG): container finished" podID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerID="9dfad846bf8620ec2edc170dc1fb8fef4c604838d0b8e64c2f6751367310fe12" exitCode=0 Oct 07 21:15:13 crc kubenswrapper[4926]: I1007 21:15:13.083017 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerDied","Data":"9dfad846bf8620ec2edc170dc1fb8fef4c604838d0b8e64c2f6751367310fe12"} Oct 07 21:15:13 crc kubenswrapper[4926]: I1007 21:15:13.084993 4926 generic.go:334] "Generic (PLEG): container finished" podID="7375476d-b601-4885-8620-4f07b35edd55" containerID="f3647224a2c7b4cf39e1d6bee3824c66704664e6234c836e868d9c7735716053" exitCode=0 Oct 07 21:15:13 crc kubenswrapper[4926]: I1007 21:15:13.085055 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-587b-account-create-csjpm" event={"ID":"7375476d-b601-4885-8620-4f07b35edd55","Type":"ContainerDied","Data":"f3647224a2c7b4cf39e1d6bee3824c66704664e6234c836e868d9c7735716053"} Oct 07 21:15:13 crc kubenswrapper[4926]: I1007 21:15:13.678998 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:15:13 crc kubenswrapper[4926]: E1007 21:15:13.680052 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:15:14 crc kubenswrapper[4926]: I1007 21:15:14.107947 4926 generic.go:334] "Generic (PLEG): container finished" podID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerID="78b4437ff7fba882a6551eddf2b75165e816cecad0c99a8d80e87bf71d03c100" exitCode=0 Oct 07 21:15:14 crc kubenswrapper[4926]: I1007 21:15:14.108024 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerDied","Data":"78b4437ff7fba882a6551eddf2b75165e816cecad0c99a8d80e87bf71d03c100"} Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.184525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-587b-account-create-csjpm" event={"ID":"7375476d-b601-4885-8620-4f07b35edd55","Type":"ContainerDied","Data":"6d411557d5ac86cfe63f36f6598e94665f069d0aa9043a8230d89b17eeb68b85"} Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.185133 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d411557d5ac86cfe63f36f6598e94665f069d0aa9043a8230d89b17eeb68b85" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.362382 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.478459 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.529109 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsdw4\" (UniqueName: \"kubernetes.io/projected/7375476d-b601-4885-8620-4f07b35edd55-kube-api-access-fsdw4\") pod \"7375476d-b601-4885-8620-4f07b35edd55\" (UID: \"7375476d-b601-4885-8620-4f07b35edd55\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.534442 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7375476d-b601-4885-8620-4f07b35edd55-kube-api-access-fsdw4" (OuterVolumeSpecName: "kube-api-access-fsdw4") pod "7375476d-b601-4885-8620-4f07b35edd55" (UID: "7375476d-b601-4885-8620-4f07b35edd55"). InnerVolumeSpecName "kube-api-access-fsdw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.630561 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-sg-core-conf-yaml\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.630937 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-scripts\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.631037 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv7n5\" (UniqueName: \"kubernetes.io/projected/24fe47d6-775e-43a2-b488-3bdd9518d4a8-kube-api-access-rv7n5\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.631080 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-combined-ca-bundle\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.631101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-log-httpd\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.631118 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-run-httpd\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.631173 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-config-data\") pod \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\" (UID: \"24fe47d6-775e-43a2-b488-3bdd9518d4a8\") " Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.631628 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsdw4\" (UniqueName: \"kubernetes.io/projected/7375476d-b601-4885-8620-4f07b35edd55-kube-api-access-fsdw4\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.635691 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-scripts" (OuterVolumeSpecName: "scripts") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.636869 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.636984 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.638287 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24fe47d6-775e-43a2-b488-3bdd9518d4a8-kube-api-access-rv7n5" (OuterVolumeSpecName: "kube-api-access-rv7n5") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "kube-api-access-rv7n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.667328 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.716080 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.737945 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.738014 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.738030 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24fe47d6-775e-43a2-b488-3bdd9518d4a8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.738043 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.738055 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.738142 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv7n5\" (UniqueName: \"kubernetes.io/projected/24fe47d6-775e-43a2-b488-3bdd9518d4a8-kube-api-access-rv7n5\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.747998 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-config-data" (OuterVolumeSpecName: "config-data") pod "24fe47d6-775e-43a2-b488-3bdd9518d4a8" (UID: "24fe47d6-775e-43a2-b488-3bdd9518d4a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:18 crc kubenswrapper[4926]: I1007 21:15:18.840799 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24fe47d6-775e-43a2-b488-3bdd9518d4a8-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.196449 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k99cm" event={"ID":"7a4ae7fe-c26e-4dca-a071-c03a34582e53","Type":"ContainerStarted","Data":"08d454e2438b30c7181ae5c3b840c1397b84367a9e4faa5c4424e26c17241dec"} Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.199991 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-587b-account-create-csjpm" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.200684 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.203360 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24fe47d6-775e-43a2-b488-3bdd9518d4a8","Type":"ContainerDied","Data":"df06be170555421c7c21b2f392545eb9f1051ba475e563ccf15dcd09d1d909b7"} Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.203431 4926 scope.go:117] "RemoveContainer" containerID="69a7fd2489d265d6a4c5d45d64be0255ddb6c4caecf44732012d6b4bf1ead8a2" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.230050 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-k99cm" podStartSLOduration=2.59175904 podStartE2EDuration="15.230030917s" podCreationTimestamp="2025-10-07 21:15:04 +0000 UTC" firstStartedPulling="2025-10-07 21:15:05.585466066 +0000 UTC m=+1155.624047216" lastFinishedPulling="2025-10-07 21:15:18.223737933 +0000 UTC m=+1168.262319093" observedRunningTime="2025-10-07 21:15:19.226995679 +0000 UTC m=+1169.265576839" watchObservedRunningTime="2025-10-07 21:15:19.230030917 +0000 UTC m=+1169.268612077" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.248209 4926 scope.go:117] "RemoveContainer" containerID="3d6ecf23f0e262653ff5e88c263648a6c8a736c25ee03c6ee904b50f2f807d83" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.251918 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.260934 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.285824 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:19 crc kubenswrapper[4926]: E1007 21:15:19.286303 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="proxy-httpd" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286323 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="proxy-httpd" Oct 07 21:15:19 crc kubenswrapper[4926]: E1007 21:15:19.286341 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-central-agent" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286351 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-central-agent" Oct 07 21:15:19 crc kubenswrapper[4926]: E1007 21:15:19.286371 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-notification-agent" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286380 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-notification-agent" Oct 07 21:15:19 crc kubenswrapper[4926]: E1007 21:15:19.286408 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7375476d-b601-4885-8620-4f07b35edd55" containerName="mariadb-account-create" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286417 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7375476d-b601-4885-8620-4f07b35edd55" containerName="mariadb-account-create" Oct 07 21:15:19 crc kubenswrapper[4926]: E1007 21:15:19.286430 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="sg-core" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286438 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="sg-core" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286668 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7375476d-b601-4885-8620-4f07b35edd55" containerName="mariadb-account-create" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286688 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-central-agent" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286704 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="ceilometer-notification-agent" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286749 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="sg-core" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.286769 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" containerName="proxy-httpd" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.288864 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.294110 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.332512 4926 scope.go:117] "RemoveContainer" containerID="9dfad846bf8620ec2edc170dc1fb8fef4c604838d0b8e64c2f6751367310fe12" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.333152 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.333247 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.365397 4926 scope.go:117] "RemoveContainer" containerID="78b4437ff7fba882a6551eddf2b75165e816cecad0c99a8d80e87bf71d03c100" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453220 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh6cs\" (UniqueName: \"kubernetes.io/projected/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-kube-api-access-jh6cs\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453292 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-scripts\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-config-data\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453348 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-log-httpd\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453383 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453454 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.453493 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-run-httpd\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.554912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-scripts\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.554973 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-config-data\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.554996 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-log-httpd\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.555031 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.555096 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.555135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-run-httpd\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.555162 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh6cs\" (UniqueName: \"kubernetes.io/projected/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-kube-api-access-jh6cs\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.555630 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-log-httpd\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.555741 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-run-httpd\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.559026 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.559943 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-config-data\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.567380 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-scripts\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.572560 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.577765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh6cs\" (UniqueName: \"kubernetes.io/projected/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-kube-api-access-jh6cs\") pod \"ceilometer-0\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " pod="openstack/ceilometer-0" Oct 07 21:15:19 crc kubenswrapper[4926]: I1007 21:15:19.650679 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:20 crc kubenswrapper[4926]: I1007 21:15:20.143602 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:20 crc kubenswrapper[4926]: I1007 21:15:20.158747 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:15:20 crc kubenswrapper[4926]: I1007 21:15:20.215039 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerStarted","Data":"93ef83a93ac8135b34aaa015ba5017a7da2ddb12dcee4fdb831ad8babdef1218"} Oct 07 21:15:20 crc kubenswrapper[4926]: I1007 21:15:20.692827 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24fe47d6-775e-43a2-b488-3bdd9518d4a8" path="/var/lib/kubelet/pods/24fe47d6-775e-43a2-b488-3bdd9518d4a8/volumes" Oct 07 21:15:21 crc kubenswrapper[4926]: I1007 21:15:21.233249 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerStarted","Data":"d7e3752b13fe3936e8e0365c7b686880fa5d7d1d15f66526ae511f917ecfd2b5"} Oct 07 21:15:21 crc kubenswrapper[4926]: I1007 21:15:21.233304 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerStarted","Data":"97b85ee794ca5f657d2a7d8a370a578ae545af88922ac43064ccf2fd1593646d"} Oct 07 21:15:22 crc kubenswrapper[4926]: E1007 21:15:22.218249 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47c8f35_5fce_4dc5_b86d_895606525d5f.slice/crio-37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69\": RecentStats: unable to find data in memory cache]" Oct 07 21:15:22 crc kubenswrapper[4926]: I1007 21:15:22.249870 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerStarted","Data":"227141c1fb1516fd4d57fabc1311e37c7f99017c11f196f56a034cb947561456"} Oct 07 21:15:24 crc kubenswrapper[4926]: I1007 21:15:24.273074 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerStarted","Data":"894660a471e00306ccd214e0df8b6a87c29e82f2c347c32af64231503d02a23a"} Oct 07 21:15:24 crc kubenswrapper[4926]: I1007 21:15:24.305392 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.956558435 podStartE2EDuration="5.305370755s" podCreationTimestamp="2025-10-07 21:15:19 +0000 UTC" firstStartedPulling="2025-10-07 21:15:20.158303081 +0000 UTC m=+1170.196884231" lastFinishedPulling="2025-10-07 21:15:23.507115371 +0000 UTC m=+1173.545696551" observedRunningTime="2025-10-07 21:15:24.299847526 +0000 UTC m=+1174.338428676" watchObservedRunningTime="2025-10-07 21:15:24.305370755 +0000 UTC m=+1174.343951905" Oct 07 21:15:25 crc kubenswrapper[4926]: I1007 21:15:25.284076 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:15:26 crc kubenswrapper[4926]: I1007 21:15:26.680505 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:15:26 crc kubenswrapper[4926]: E1007 21:15:26.680787 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(392cbe25-f961-43ce-bc58-453dba0de3a9)\"" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" Oct 07 21:15:32 crc kubenswrapper[4926]: E1007 21:15:32.503461 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47c8f35_5fce_4dc5_b86d_895606525d5f.slice/crio-37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69\": RecentStats: unable to find data in memory cache]" Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.387031 4926 generic.go:334] "Generic (PLEG): container finished" podID="7a4ae7fe-c26e-4dca-a071-c03a34582e53" containerID="08d454e2438b30c7181ae5c3b840c1397b84367a9e4faa5c4424e26c17241dec" exitCode=0 Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.387184 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k99cm" event={"ID":"7a4ae7fe-c26e-4dca-a071-c03a34582e53","Type":"ContainerDied","Data":"08d454e2438b30c7181ae5c3b840c1397b84367a9e4faa5c4424e26c17241dec"} Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.460299 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.460736 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-central-agent" containerID="cri-o://97b85ee794ca5f657d2a7d8a370a578ae545af88922ac43064ccf2fd1593646d" gracePeriod=30 Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.460882 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="proxy-httpd" containerID="cri-o://894660a471e00306ccd214e0df8b6a87c29e82f2c347c32af64231503d02a23a" gracePeriod=30 Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.460916 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="sg-core" containerID="cri-o://227141c1fb1516fd4d57fabc1311e37c7f99017c11f196f56a034cb947561456" gracePeriod=30 Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.461006 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-notification-agent" containerID="cri-o://d7e3752b13fe3936e8e0365c7b686880fa5d7d1d15f66526ae511f917ecfd2b5" gracePeriod=30 Oct 07 21:15:33 crc kubenswrapper[4926]: I1007 21:15:33.474723 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.202:3000/\": EOF" Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.402737 4926 generic.go:334] "Generic (PLEG): container finished" podID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerID="894660a471e00306ccd214e0df8b6a87c29e82f2c347c32af64231503d02a23a" exitCode=0 Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.403042 4926 generic.go:334] "Generic (PLEG): container finished" podID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerID="227141c1fb1516fd4d57fabc1311e37c7f99017c11f196f56a034cb947561456" exitCode=2 Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.403055 4926 generic.go:334] "Generic (PLEG): container finished" podID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerID="97b85ee794ca5f657d2a7d8a370a578ae545af88922ac43064ccf2fd1593646d" exitCode=0 Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.402801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerDied","Data":"894660a471e00306ccd214e0df8b6a87c29e82f2c347c32af64231503d02a23a"} Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.403132 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerDied","Data":"227141c1fb1516fd4d57fabc1311e37c7f99017c11f196f56a034cb947561456"} Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.403175 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerDied","Data":"97b85ee794ca5f657d2a7d8a370a578ae545af88922ac43064ccf2fd1593646d"} Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.801473 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.816128 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-scripts\") pod \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.825144 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-scripts" (OuterVolumeSpecName: "scripts") pod "7a4ae7fe-c26e-4dca-a071-c03a34582e53" (UID: "7a4ae7fe-c26e-4dca-a071-c03a34582e53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.918920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcvgb\" (UniqueName: \"kubernetes.io/projected/7a4ae7fe-c26e-4dca-a071-c03a34582e53-kube-api-access-wcvgb\") pod \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.919221 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-combined-ca-bundle\") pod \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.919262 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-config-data\") pod \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\" (UID: \"7a4ae7fe-c26e-4dca-a071-c03a34582e53\") " Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.920012 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.928578 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a4ae7fe-c26e-4dca-a071-c03a34582e53-kube-api-access-wcvgb" (OuterVolumeSpecName: "kube-api-access-wcvgb") pod "7a4ae7fe-c26e-4dca-a071-c03a34582e53" (UID: "7a4ae7fe-c26e-4dca-a071-c03a34582e53"). InnerVolumeSpecName "kube-api-access-wcvgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.964767 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a4ae7fe-c26e-4dca-a071-c03a34582e53" (UID: "7a4ae7fe-c26e-4dca-a071-c03a34582e53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:34 crc kubenswrapper[4926]: I1007 21:15:34.974157 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-config-data" (OuterVolumeSpecName: "config-data") pod "7a4ae7fe-c26e-4dca-a071-c03a34582e53" (UID: "7a4ae7fe-c26e-4dca-a071-c03a34582e53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.022446 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcvgb\" (UniqueName: \"kubernetes.io/projected/7a4ae7fe-c26e-4dca-a071-c03a34582e53-kube-api-access-wcvgb\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.022507 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.022529 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4ae7fe-c26e-4dca-a071-c03a34582e53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.418040 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k99cm" event={"ID":"7a4ae7fe-c26e-4dca-a071-c03a34582e53","Type":"ContainerDied","Data":"c60cb4c34c03fea90a81a1a96b166082a0dc1eb9d680a8e7218ae80c3733fac4"} Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.418456 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c60cb4c34c03fea90a81a1a96b166082a0dc1eb9d680a8e7218ae80c3733fac4" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.418157 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k99cm" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.542615 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 21:15:35 crc kubenswrapper[4926]: E1007 21:15:35.543442 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a4ae7fe-c26e-4dca-a071-c03a34582e53" containerName="nova-cell0-conductor-db-sync" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.543553 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a4ae7fe-c26e-4dca-a071-c03a34582e53" containerName="nova-cell0-conductor-db-sync" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.543911 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a4ae7fe-c26e-4dca-a071-c03a34582e53" containerName="nova-cell0-conductor-db-sync" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.544876 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.547574 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5qfg2" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.549530 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.556605 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.634510 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4601f7d-946c-4632-9778-44f70cafc757-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.634620 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pv2t\" (UniqueName: \"kubernetes.io/projected/b4601f7d-946c-4632-9778-44f70cafc757-kube-api-access-6pv2t\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.634800 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4601f7d-946c-4632-9778-44f70cafc757-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.736321 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pv2t\" (UniqueName: \"kubernetes.io/projected/b4601f7d-946c-4632-9778-44f70cafc757-kube-api-access-6pv2t\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.737121 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4601f7d-946c-4632-9778-44f70cafc757-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.737184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4601f7d-946c-4632-9778-44f70cafc757-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.741629 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4601f7d-946c-4632-9778-44f70cafc757-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.744393 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4601f7d-946c-4632-9778-44f70cafc757-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.753010 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pv2t\" (UniqueName: \"kubernetes.io/projected/b4601f7d-946c-4632-9778-44f70cafc757-kube-api-access-6pv2t\") pod \"nova-cell0-conductor-0\" (UID: \"b4601f7d-946c-4632-9778-44f70cafc757\") " pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:35 crc kubenswrapper[4926]: I1007 21:15:35.868550 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.378930 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 21:15:36 crc kubenswrapper[4926]: W1007 21:15:36.403466 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4601f7d_946c_4632_9778_44f70cafc757.slice/crio-d54515340cdd148fac029550f055489ae636e1fb6b4d562158eae5f7a2eafa73 WatchSource:0}: Error finding container d54515340cdd148fac029550f055489ae636e1fb6b4d562158eae5f7a2eafa73: Status 404 returned error can't find the container with id d54515340cdd148fac029550f055489ae636e1fb6b4d562158eae5f7a2eafa73 Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.435976 4926 generic.go:334] "Generic (PLEG): container finished" podID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerID="d7e3752b13fe3936e8e0365c7b686880fa5d7d1d15f66526ae511f917ecfd2b5" exitCode=0 Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.436181 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerDied","Data":"d7e3752b13fe3936e8e0365c7b686880fa5d7d1d15f66526ae511f917ecfd2b5"} Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.437994 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b4601f7d-946c-4632-9778-44f70cafc757","Type":"ContainerStarted","Data":"d54515340cdd148fac029550f055489ae636e1fb6b4d562158eae5f7a2eafa73"} Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.735485 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760159 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-config-data\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760294 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-sg-core-conf-yaml\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760369 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-scripts\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760519 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh6cs\" (UniqueName: \"kubernetes.io/projected/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-kube-api-access-jh6cs\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760573 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-combined-ca-bundle\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760617 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-log-httpd\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.760667 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-run-httpd\") pod \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\" (UID: \"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e\") " Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.761274 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.761337 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.762836 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.762899 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.765091 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-kube-api-access-jh6cs" (OuterVolumeSpecName: "kube-api-access-jh6cs") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "kube-api-access-jh6cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.767611 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-scripts" (OuterVolumeSpecName: "scripts") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.803024 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.841659 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.865424 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.865599 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh6cs\" (UniqueName: \"kubernetes.io/projected/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-kube-api-access-jh6cs\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.865683 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.865756 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.873358 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-config-data" (OuterVolumeSpecName: "config-data") pod "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" (UID: "fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:36 crc kubenswrapper[4926]: I1007 21:15:36.970065 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.452758 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e","Type":"ContainerDied","Data":"93ef83a93ac8135b34aaa015ba5017a7da2ddb12dcee4fdb831ad8babdef1218"} Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.453045 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.453058 4926 scope.go:117] "RemoveContainer" containerID="894660a471e00306ccd214e0df8b6a87c29e82f2c347c32af64231503d02a23a" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.456414 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b4601f7d-946c-4632-9778-44f70cafc757","Type":"ContainerStarted","Data":"93b04272fd458b4f41b91ef5f74342bce3dcba924eca5449c027f0f19c7d7f1f"} Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.457676 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.494476 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.494451529 podStartE2EDuration="2.494451529s" podCreationTimestamp="2025-10-07 21:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:15:37.483301557 +0000 UTC m=+1187.521882707" watchObservedRunningTime="2025-10-07 21:15:37.494451529 +0000 UTC m=+1187.533032689" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.503861 4926 scope.go:117] "RemoveContainer" containerID="227141c1fb1516fd4d57fabc1311e37c7f99017c11f196f56a034cb947561456" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.506503 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.517809 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.529605 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:37 crc kubenswrapper[4926]: E1007 21:15:37.530230 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="proxy-httpd" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.530324 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="proxy-httpd" Oct 07 21:15:37 crc kubenswrapper[4926]: E1007 21:15:37.530383 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="sg-core" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.530440 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="sg-core" Oct 07 21:15:37 crc kubenswrapper[4926]: E1007 21:15:37.530497 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-central-agent" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.530554 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-central-agent" Oct 07 21:15:37 crc kubenswrapper[4926]: E1007 21:15:37.530608 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-notification-agent" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.530662 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-notification-agent" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.530870 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="proxy-httpd" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.530947 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-central-agent" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.531109 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="ceilometer-notification-agent" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.531222 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" containerName="sg-core" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.532874 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.550558 4926 scope.go:117] "RemoveContainer" containerID="d7e3752b13fe3936e8e0365c7b686880fa5d7d1d15f66526ae511f917ecfd2b5" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.565760 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.574667 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.584743 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-config-data\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.584936 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-run-httpd\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.585104 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-log-httpd\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.585241 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbj45\" (UniqueName: \"kubernetes.io/projected/000382f4-d2d7-4c63-aa0b-e14c64dca78c-kube-api-access-bbj45\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.585379 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.585454 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-scripts\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.585541 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.591303 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.674364 4926 scope.go:117] "RemoveContainer" containerID="97b85ee794ca5f657d2a7d8a370a578ae545af88922ac43064ccf2fd1593646d" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688668 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-scripts\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688786 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-config-data\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688823 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-run-httpd\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-log-httpd\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.688918 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbj45\" (UniqueName: \"kubernetes.io/projected/000382f4-d2d7-4c63-aa0b-e14c64dca78c-kube-api-access-bbj45\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.690799 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-run-httpd\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.691100 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-log-httpd\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.700898 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.703834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-scripts\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.706633 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.706941 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-config-data\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.722853 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbj45\" (UniqueName: \"kubernetes.io/projected/000382f4-d2d7-4c63-aa0b-e14c64dca78c-kube-api-access-bbj45\") pod \"ceilometer-0\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " pod="openstack/ceilometer-0" Oct 07 21:15:37 crc kubenswrapper[4926]: I1007 21:15:37.867750 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:15:38 crc kubenswrapper[4926]: I1007 21:15:38.354536 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:15:38 crc kubenswrapper[4926]: I1007 21:15:38.467870 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerStarted","Data":"2b6a5fcecdb0fcef18200cdc386e0007f6c685ea584d72df7e34e611a4ca3e38"} Oct 07 21:15:38 crc kubenswrapper[4926]: I1007 21:15:38.688350 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e" path="/var/lib/kubelet/pods/fbc05fc8-4acc-4b69-b2e7-79b3f2d5798e/volumes" Oct 07 21:15:39 crc kubenswrapper[4926]: I1007 21:15:39.485340 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerStarted","Data":"8b34a8a9db2d6f0b0378f2fdb5c799275b840364fa246ed26d0c9c6740427741"} Oct 07 21:15:39 crc kubenswrapper[4926]: I1007 21:15:39.485658 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerStarted","Data":"658823b71079c86298e5a27d8a6eb88cd783ea6c2d3a65a0fbc934a4e52a44f3"} Oct 07 21:15:40 crc kubenswrapper[4926]: I1007 21:15:40.500734 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerStarted","Data":"9f6d94f99d0ac2720efb728ceb1536625581eb30845ded5a9cb5483ecb6a4508"} Oct 07 21:15:40 crc kubenswrapper[4926]: I1007 21:15:40.710385 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:15:41 crc kubenswrapper[4926]: I1007 21:15:41.516705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerStarted","Data":"8b4ddf88fe6eb7a6f2176d43fffd98a3c21e616d423be53c2f6066cc3b0a5465"} Oct 07 21:15:41 crc kubenswrapper[4926]: I1007 21:15:41.517239 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:15:41 crc kubenswrapper[4926]: I1007 21:15:41.519727 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerStarted","Data":"eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5"} Oct 07 21:15:41 crc kubenswrapper[4926]: I1007 21:15:41.539951 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.751372742 podStartE2EDuration="4.539929273s" podCreationTimestamp="2025-10-07 21:15:37 +0000 UTC" firstStartedPulling="2025-10-07 21:15:38.372771842 +0000 UTC m=+1188.411353022" lastFinishedPulling="2025-10-07 21:15:41.161328393 +0000 UTC m=+1191.199909553" observedRunningTime="2025-10-07 21:15:41.538829281 +0000 UTC m=+1191.577410441" watchObservedRunningTime="2025-10-07 21:15:41.539929273 +0000 UTC m=+1191.578510433" Oct 07 21:15:42 crc kubenswrapper[4926]: E1007 21:15:42.840934 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47c8f35_5fce_4dc5_b86d_895606525d5f.slice/crio-37577da59e0cca2188b4a59ba4fdf274b61aa534ccad87c9abce48f264221b69\": RecentStats: unable to find data in memory cache]" Oct 07 21:15:45 crc kubenswrapper[4926]: I1007 21:15:45.915136 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.449910 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zl2hw"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.451394 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.455128 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.455468 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.480782 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zl2hw"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.481431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.481509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-config-data\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.481560 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lq47\" (UniqueName: \"kubernetes.io/projected/a505eae2-6bc8-40cd-a133-9bd9bcb98478-kube-api-access-5lq47\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.481639 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-scripts\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.583525 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.583561 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-config-data\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.583599 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lq47\" (UniqueName: \"kubernetes.io/projected/a505eae2-6bc8-40cd-a133-9bd9bcb98478-kube-api-access-5lq47\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.583892 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-scripts\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.594876 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.595872 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-config-data\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.618765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-scripts\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.635904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lq47\" (UniqueName: \"kubernetes.io/projected/a505eae2-6bc8-40cd-a133-9bd9bcb98478-kube-api-access-5lq47\") pod \"nova-cell0-cell-mapping-zl2hw\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.783805 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.820681 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.824217 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.833081 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.838335 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.866266 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.867910 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.871516 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.908748 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944394 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944452 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-config-data\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944553 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm6b7\" (UniqueName: \"kubernetes.io/projected/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-kube-api-access-rm6b7\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944601 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-logs\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944654 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944675 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.944705 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjlz2\" (UniqueName: \"kubernetes.io/projected/87abc81b-808f-4752-8ae9-5e34e609138b-kube-api-access-kjlz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.974258 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.986661 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:15:46 crc kubenswrapper[4926]: I1007 21:15:46.988505 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.033382 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046635 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-logs\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046731 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046758 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzt8r\" (UniqueName: \"kubernetes.io/projected/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-kube-api-access-fzt8r\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046814 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjlz2\" (UniqueName: \"kubernetes.io/projected/87abc81b-808f-4752-8ae9-5e34e609138b-kube-api-access-kjlz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046834 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046855 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-config-data\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046898 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046916 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-logs\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.046979 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-config-data\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.047012 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm6b7\" (UniqueName: \"kubernetes.io/projected/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-kube-api-access-rm6b7\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.047658 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-logs\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.056271 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8685f8ddff-w54pw"] Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.060569 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.060756 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-config-data\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.067753 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.068978 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.071574 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.086315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm6b7\" (UniqueName: \"kubernetes.io/projected/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-kube-api-access-rm6b7\") pod \"nova-metadata-0\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.109411 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8685f8ddff-w54pw"] Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.110746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjlz2\" (UniqueName: \"kubernetes.io/projected/87abc81b-808f-4752-8ae9-5e34e609138b-kube-api-access-kjlz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.124421 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.125709 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.127522 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.137256 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148488 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-nb\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzt8r\" (UniqueName: \"kubernetes.io/projected/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-kube-api-access-fzt8r\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148600 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh67s\" (UniqueName: \"kubernetes.io/projected/15217f6a-e2ad-4d8d-b675-e3aec437574f-kube-api-access-dh67s\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-swift-storage-0\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148659 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-svc\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148688 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltm58\" (UniqueName: \"kubernetes.io/projected/3cf26616-4656-4a96-b782-fb17eb11b912-kube-api-access-ltm58\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148709 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-sb\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148778 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148801 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-logs\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148853 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148903 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-config-data\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148929 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-config-data\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.148960 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-config\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.150635 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-logs\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.165881 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.166005 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-config-data\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.170381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzt8r\" (UniqueName: \"kubernetes.io/projected/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-kube-api-access-fzt8r\") pod \"nova-api-0\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.250441 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-nb\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.250517 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh67s\" (UniqueName: \"kubernetes.io/projected/15217f6a-e2ad-4d8d-b675-e3aec437574f-kube-api-access-dh67s\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.250543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-swift-storage-0\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.250561 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-svc\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.250579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltm58\" (UniqueName: \"kubernetes.io/projected/3cf26616-4656-4a96-b782-fb17eb11b912-kube-api-access-ltm58\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.251482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-sb\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.251516 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-sb\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.251619 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.251658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-config-data\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.251685 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-config\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.252287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-config\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.252700 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-swift-storage-0\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.253718 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-svc\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.259838 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-nb\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.259890 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.262796 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-config-data\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.273847 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh67s\" (UniqueName: \"kubernetes.io/projected/15217f6a-e2ad-4d8d-b675-e3aec437574f-kube-api-access-dh67s\") pod \"nova-scheduler-0\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.276244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltm58\" (UniqueName: \"kubernetes.io/projected/3cf26616-4656-4a96-b782-fb17eb11b912-kube-api-access-ltm58\") pod \"dnsmasq-dns-8685f8ddff-w54pw\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.288624 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.327692 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.430168 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.467071 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.475717 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.500772 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zl2hw"] Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.631441 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zl2hw" event={"ID":"a505eae2-6bc8-40cd-a133-9bd9bcb98478","Type":"ContainerStarted","Data":"d103be10cbf04467dc1536d843697847311af7931411c94c6b43789dc5b860db"} Oct 07 21:15:47 crc kubenswrapper[4926]: I1007 21:15:47.888674 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:47.998038 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.010752 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbgql"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.012096 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: W1007 21:15:48.013368 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff0e6f9_e423_46fe_835c_ee19b9f302ac.slice/crio-34c2aa91098d93c0d7356e49c894e92216decb3262744e4a9eed8d5160a85a17 WatchSource:0}: Error finding container 34c2aa91098d93c0d7356e49c894e92216decb3262744e4a9eed8d5160a85a17: Status 404 returned error can't find the container with id 34c2aa91098d93c0d7356e49c894e92216decb3262744e4a9eed8d5160a85a17 Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.015302 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.015689 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.018675 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbgql"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.075833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-config-data\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.076065 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.076100 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtccv\" (UniqueName: \"kubernetes.io/projected/111e3572-9e8a-4964-bc03-c5a50581e0f7-kube-api-access-dtccv\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.076175 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-scripts\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.177931 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8685f8ddff-w54pw"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.178818 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtccv\" (UniqueName: \"kubernetes.io/projected/111e3572-9e8a-4964-bc03-c5a50581e0f7-kube-api-access-dtccv\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.178872 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.178932 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-scripts\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.179070 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-config-data\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: W1007 21:15:48.183917 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cf26616_4656_4a96_b782_fb17eb11b912.slice/crio-7865f89c3c3c2175990d1a7d3dc742f53b1719988ea18c17237475c6185b65f2 WatchSource:0}: Error finding container 7865f89c3c3c2175990d1a7d3dc742f53b1719988ea18c17237475c6185b65f2: Status 404 returned error can't find the container with id 7865f89c3c3c2175990d1a7d3dc742f53b1719988ea18c17237475c6185b65f2 Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.186555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-config-data\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.186938 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-scripts\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.199635 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtccv\" (UniqueName: \"kubernetes.io/projected/111e3572-9e8a-4964-bc03-c5a50581e0f7-kube-api-access-dtccv\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.203478 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vbgql\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.217185 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.322026 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.333891 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.705240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82","Type":"ContainerStarted","Data":"e1a76134948a8486a11f46b43aebf127d0785488d5ada2f3a402c5c7b098e578"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.728479 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"87abc81b-808f-4752-8ae9-5e34e609138b","Type":"ContainerStarted","Data":"7c15b9d613cbc1bebb488aa4622b7be2bf0ab5939c1eaed3079c8d5cf04f9c20"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.747519 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zl2hw" event={"ID":"a505eae2-6bc8-40cd-a133-9bd9bcb98478","Type":"ContainerStarted","Data":"f90f06b4e30afb44c9fdcf03e35e8aae59be7ac3b3da3be2ccf994e0944b8ccb"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.761471 4926 generic.go:334] "Generic (PLEG): container finished" podID="3cf26616-4656-4a96-b782-fb17eb11b912" containerID="cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8" exitCode=0 Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.761788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" event={"ID":"3cf26616-4656-4a96-b782-fb17eb11b912","Type":"ContainerDied","Data":"cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.761851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" event={"ID":"3cf26616-4656-4a96-b782-fb17eb11b912","Type":"ContainerStarted","Data":"7865f89c3c3c2175990d1a7d3dc742f53b1719988ea18c17237475c6185b65f2"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.771146 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15217f6a-e2ad-4d8d-b675-e3aec437574f","Type":"ContainerStarted","Data":"7da87950545703d3308029067e58f7dbe43afdfce2f9fe2eb4a0602efade4e47"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.775597 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zl2hw" podStartSLOduration=2.775576921 podStartE2EDuration="2.775576921s" podCreationTimestamp="2025-10-07 21:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:15:48.771085961 +0000 UTC m=+1198.809667111" watchObservedRunningTime="2025-10-07 21:15:48.775576921 +0000 UTC m=+1198.814158071" Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.778815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ff0e6f9-e423-46fe-835c-ee19b9f302ac","Type":"ContainerStarted","Data":"34c2aa91098d93c0d7356e49c894e92216decb3262744e4a9eed8d5160a85a17"} Oct 07 21:15:48 crc kubenswrapper[4926]: I1007 21:15:48.882858 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbgql"] Oct 07 21:15:50 crc kubenswrapper[4926]: W1007 21:15:50.332264 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111e3572_9e8a_4964_bc03_c5a50581e0f7.slice/crio-fdd76e171ec5d0ed86e8487828b7f1133eed6c61f4fd96d55e9bad4289982ca9 WatchSource:0}: Error finding container fdd76e171ec5d0ed86e8487828b7f1133eed6c61f4fd96d55e9bad4289982ca9: Status 404 returned error can't find the container with id fdd76e171ec5d0ed86e8487828b7f1133eed6c61f4fd96d55e9bad4289982ca9 Oct 07 21:15:50 crc kubenswrapper[4926]: I1007 21:15:50.724473 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:15:50 crc kubenswrapper[4926]: I1007 21:15:50.724516 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:50 crc kubenswrapper[4926]: I1007 21:15:50.797543 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbgql" event={"ID":"111e3572-9e8a-4964-bc03-c5a50581e0f7","Type":"ContainerStarted","Data":"fdd76e171ec5d0ed86e8487828b7f1133eed6c61f4fd96d55e9bad4289982ca9"} Oct 07 21:15:50 crc kubenswrapper[4926]: I1007 21:15:50.946658 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:50 crc kubenswrapper[4926]: I1007 21:15:50.978902 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.853384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15217f6a-e2ad-4d8d-b675-e3aec437574f","Type":"ContainerStarted","Data":"7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25"} Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.878552 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.027171599 podStartE2EDuration="5.87853611s" podCreationTimestamp="2025-10-07 21:15:46 +0000 UTC" firstStartedPulling="2025-10-07 21:15:48.362711833 +0000 UTC m=+1198.401292983" lastFinishedPulling="2025-10-07 21:15:51.214076344 +0000 UTC m=+1201.252657494" observedRunningTime="2025-10-07 21:15:51.876319426 +0000 UTC m=+1201.914900576" watchObservedRunningTime="2025-10-07 21:15:51.87853611 +0000 UTC m=+1201.917117250" Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.896583 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82","Type":"ContainerStarted","Data":"429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a"} Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.903459 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"87abc81b-808f-4752-8ae9-5e34e609138b","Type":"ContainerStarted","Data":"4add980846fd6d10bc5891c0d57617d4f4c9ab3f7d342059614a29700e4e918d"} Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.904003 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="87abc81b-808f-4752-8ae9-5e34e609138b" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4add980846fd6d10bc5891c0d57617d4f4c9ab3f7d342059614a29700e4e918d" gracePeriod=30 Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.909502 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbgql" event={"ID":"111e3572-9e8a-4964-bc03-c5a50581e0f7","Type":"ContainerStarted","Data":"379872a104dbc4211471dfe06bd17267e6e5c5c3cbb195aaa20b1e77d143e971"} Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.909558 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.934663 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.62956709 podStartE2EDuration="5.934637008s" podCreationTimestamp="2025-10-07 21:15:46 +0000 UTC" firstStartedPulling="2025-10-07 21:15:47.907426191 +0000 UTC m=+1197.946007341" lastFinishedPulling="2025-10-07 21:15:51.212496109 +0000 UTC m=+1201.251077259" observedRunningTime="2025-10-07 21:15:51.919052658 +0000 UTC m=+1201.957633808" watchObservedRunningTime="2025-10-07 21:15:51.934637008 +0000 UTC m=+1201.973218158" Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.950551 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vbgql" podStartSLOduration=4.950509436 podStartE2EDuration="4.950509436s" podCreationTimestamp="2025-10-07 21:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:15:51.942420882 +0000 UTC m=+1201.981002032" watchObservedRunningTime="2025-10-07 21:15:51.950509436 +0000 UTC m=+1201.989090586" Oct 07 21:15:51 crc kubenswrapper[4926]: I1007 21:15:51.977386 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.015681 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.288987 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.477274 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.934925 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82","Type":"ContainerStarted","Data":"a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15"} Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.946104 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" event={"ID":"3cf26616-4656-4a96-b782-fb17eb11b912","Type":"ContainerStarted","Data":"a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da"} Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.947052 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.950309 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-log" containerID="cri-o://44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08" gracePeriod=30 Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.950723 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-metadata" containerID="cri-o://d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e" gracePeriod=30 Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.950442 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ff0e6f9-e423-46fe-835c-ee19b9f302ac","Type":"ContainerStarted","Data":"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08"} Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.954900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ff0e6f9-e423-46fe-835c-ee19b9f302ac","Type":"ContainerStarted","Data":"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e"} Oct 07 21:15:52 crc kubenswrapper[4926]: I1007 21:15:52.977257 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.982243856 podStartE2EDuration="6.97723109s" podCreationTimestamp="2025-10-07 21:15:46 +0000 UTC" firstStartedPulling="2025-10-07 21:15:48.217527145 +0000 UTC m=+1198.256108295" lastFinishedPulling="2025-10-07 21:15:51.212514379 +0000 UTC m=+1201.251095529" observedRunningTime="2025-10-07 21:15:52.960970481 +0000 UTC m=+1202.999551651" watchObservedRunningTime="2025-10-07 21:15:52.97723109 +0000 UTC m=+1203.015812260" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.024839 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.8275640539999998 podStartE2EDuration="7.024799312s" podCreationTimestamp="2025-10-07 21:15:46 +0000 UTC" firstStartedPulling="2025-10-07 21:15:48.016929789 +0000 UTC m=+1198.055510939" lastFinishedPulling="2025-10-07 21:15:51.214165047 +0000 UTC m=+1201.252746197" observedRunningTime="2025-10-07 21:15:53.002501449 +0000 UTC m=+1203.041082599" watchObservedRunningTime="2025-10-07 21:15:53.024799312 +0000 UTC m=+1203.063380482" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.028674 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" podStartSLOduration=7.028665193 podStartE2EDuration="7.028665193s" podCreationTimestamp="2025-10-07 21:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:15:53.026453489 +0000 UTC m=+1203.065034629" watchObservedRunningTime="2025-10-07 21:15:53.028665193 +0000 UTC m=+1203.067246343" Oct 07 21:15:53 crc kubenswrapper[4926]: E1007 21:15:53.117647 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff0e6f9_e423_46fe_835c_ee19b9f302ac.slice/crio-conmon-44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08.scope\": RecentStats: unable to find data in memory cache]" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.580746 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.625566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-combined-ca-bundle\") pod \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.625747 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm6b7\" (UniqueName: \"kubernetes.io/projected/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-kube-api-access-rm6b7\") pod \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.625844 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-config-data\") pod \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.625916 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-logs\") pod \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\" (UID: \"2ff0e6f9-e423-46fe-835c-ee19b9f302ac\") " Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.626768 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-logs" (OuterVolumeSpecName: "logs") pod "2ff0e6f9-e423-46fe-835c-ee19b9f302ac" (UID: "2ff0e6f9-e423-46fe-835c-ee19b9f302ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.631972 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-kube-api-access-rm6b7" (OuterVolumeSpecName: "kube-api-access-rm6b7") pod "2ff0e6f9-e423-46fe-835c-ee19b9f302ac" (UID: "2ff0e6f9-e423-46fe-835c-ee19b9f302ac"). InnerVolumeSpecName "kube-api-access-rm6b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.658833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-config-data" (OuterVolumeSpecName: "config-data") pod "2ff0e6f9-e423-46fe-835c-ee19b9f302ac" (UID: "2ff0e6f9-e423-46fe-835c-ee19b9f302ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.673495 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ff0e6f9-e423-46fe-835c-ee19b9f302ac" (UID: "2ff0e6f9-e423-46fe-835c-ee19b9f302ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.728041 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.728078 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm6b7\" (UniqueName: \"kubernetes.io/projected/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-kube-api-access-rm6b7\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.728090 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.728099 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff0e6f9-e423-46fe-835c-ee19b9f302ac-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960077 4926 generic.go:334] "Generic (PLEG): container finished" podID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerID="d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e" exitCode=0 Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960450 4926 generic.go:334] "Generic (PLEG): container finished" podID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerID="44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08" exitCode=143 Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960653 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" containerID="cri-o://eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5" gracePeriod=30 Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960304 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ff0e6f9-e423-46fe-835c-ee19b9f302ac","Type":"ContainerDied","Data":"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e"} Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960740 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ff0e6f9-e423-46fe-835c-ee19b9f302ac","Type":"ContainerDied","Data":"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08"} Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ff0e6f9-e423-46fe-835c-ee19b9f302ac","Type":"ContainerDied","Data":"34c2aa91098d93c0d7356e49c894e92216decb3262744e4a9eed8d5160a85a17"} Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960771 4926 scope.go:117] "RemoveContainer" containerID="d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.960368 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.984345 4926 scope.go:117] "RemoveContainer" containerID="44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08" Oct 07 21:15:53 crc kubenswrapper[4926]: I1007 21:15:53.993330 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.001966 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.017899 4926 scope.go:117] "RemoveContainer" containerID="d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e" Oct 07 21:15:54 crc kubenswrapper[4926]: E1007 21:15:54.020073 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e\": container with ID starting with d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e not found: ID does not exist" containerID="d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.020104 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e"} err="failed to get container status \"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e\": rpc error: code = NotFound desc = could not find container \"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e\": container with ID starting with d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e not found: ID does not exist" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.020125 4926 scope.go:117] "RemoveContainer" containerID="44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08" Oct 07 21:15:54 crc kubenswrapper[4926]: E1007 21:15:54.021307 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08\": container with ID starting with 44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08 not found: ID does not exist" containerID="44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.021333 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08"} err="failed to get container status \"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08\": rpc error: code = NotFound desc = could not find container \"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08\": container with ID starting with 44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08 not found: ID does not exist" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.021347 4926 scope.go:117] "RemoveContainer" containerID="d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.022781 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e"} err="failed to get container status \"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e\": rpc error: code = NotFound desc = could not find container \"d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e\": container with ID starting with d1d8ecd8ce0d4419c7192b021c21ddb9c4d0804964562e75e73f28761019fc3e not found: ID does not exist" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.022802 4926 scope.go:117] "RemoveContainer" containerID="44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.025596 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08"} err="failed to get container status \"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08\": rpc error: code = NotFound desc = could not find container \"44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08\": container with ID starting with 44f5adf5154472ee0fa4838b3cce4839c396eb722e2f8d484ac4785676dd3e08 not found: ID does not exist" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.029206 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:54 crc kubenswrapper[4926]: E1007 21:15:54.029637 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-metadata" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.029653 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-metadata" Oct 07 21:15:54 crc kubenswrapper[4926]: E1007 21:15:54.029667 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-log" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.029674 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-log" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.029854 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-metadata" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.029875 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" containerName="nova-metadata-log" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.030912 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.033305 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.036468 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.046508 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.135378 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.135665 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-config-data\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.135781 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f2b397-8462-4fbf-a88f-40a4c803c3d0-logs\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.135908 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.136030 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6wpc\" (UniqueName: \"kubernetes.io/projected/68f2b397-8462-4fbf-a88f-40a4c803c3d0-kube-api-access-f6wpc\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.237729 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.237841 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6wpc\" (UniqueName: \"kubernetes.io/projected/68f2b397-8462-4fbf-a88f-40a4c803c3d0-kube-api-access-f6wpc\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.237928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.237954 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-config-data\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.237983 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f2b397-8462-4fbf-a88f-40a4c803c3d0-logs\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.238431 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f2b397-8462-4fbf-a88f-40a4c803c3d0-logs\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.242396 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.242545 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-config-data\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.261827 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.280792 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6wpc\" (UniqueName: \"kubernetes.io/projected/68f2b397-8462-4fbf-a88f-40a4c803c3d0-kube-api-access-f6wpc\") pod \"nova-metadata-0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.371682 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.692886 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff0e6f9-e423-46fe-835c-ee19b9f302ac" path="/var/lib/kubelet/pods/2ff0e6f9-e423-46fe-835c-ee19b9f302ac/volumes" Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.826936 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:15:54 crc kubenswrapper[4926]: I1007 21:15:54.975896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f2b397-8462-4fbf-a88f-40a4c803c3d0","Type":"ContainerStarted","Data":"dc9c2aeee3bfbceae7eeb33effa556b0a94f8e97cbd251e2eb12ae93e943b4dc"} Oct 07 21:15:55 crc kubenswrapper[4926]: I1007 21:15:55.994037 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f2b397-8462-4fbf-a88f-40a4c803c3d0","Type":"ContainerStarted","Data":"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916"} Oct 07 21:15:55 crc kubenswrapper[4926]: I1007 21:15:55.994645 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f2b397-8462-4fbf-a88f-40a4c803c3d0","Type":"ContainerStarted","Data":"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4"} Oct 07 21:15:56 crc kubenswrapper[4926]: I1007 21:15:56.019776 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.019752075 podStartE2EDuration="3.019752075s" podCreationTimestamp="2025-10-07 21:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:15:56.019738515 +0000 UTC m=+1206.058319675" watchObservedRunningTime="2025-10-07 21:15:56.019752075 +0000 UTC m=+1206.058333245" Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.431389 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.431799 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.469817 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.476447 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.543173 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.551107 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84c5b587c5-jvms2"] Oct 07 21:15:57 crc kubenswrapper[4926]: I1007 21:15:57.551355 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" containerName="dnsmasq-dns" containerID="cri-o://ad0f1088b7fafad316bdcc107a5ed82e953a65ca275ebaf97ec7db01a4518852" gracePeriod=10 Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.038703 4926 generic.go:334] "Generic (PLEG): container finished" podID="a772eddc-1462-4636-a4ba-1340b24dc102" containerID="ad0f1088b7fafad316bdcc107a5ed82e953a65ca275ebaf97ec7db01a4518852" exitCode=0 Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.038770 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" event={"ID":"a772eddc-1462-4636-a4ba-1340b24dc102","Type":"ContainerDied","Data":"ad0f1088b7fafad316bdcc107a5ed82e953a65ca275ebaf97ec7db01a4518852"} Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.040171 4926 generic.go:334] "Generic (PLEG): container finished" podID="a505eae2-6bc8-40cd-a133-9bd9bcb98478" containerID="f90f06b4e30afb44c9fdcf03e35e8aae59be7ac3b3da3be2ccf994e0944b8ccb" exitCode=0 Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.040350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zl2hw" event={"ID":"a505eae2-6bc8-40cd-a133-9bd9bcb98478","Type":"ContainerDied","Data":"f90f06b4e30afb44c9fdcf03e35e8aae59be7ac3b3da3be2ccf994e0944b8ccb"} Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.099033 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.161533 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.336732 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-sb\") pod \"a772eddc-1462-4636-a4ba-1340b24dc102\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.336782 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d872c\" (UniqueName: \"kubernetes.io/projected/a772eddc-1462-4636-a4ba-1340b24dc102-kube-api-access-d872c\") pod \"a772eddc-1462-4636-a4ba-1340b24dc102\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.336880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-nb\") pod \"a772eddc-1462-4636-a4ba-1340b24dc102\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.336942 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-config\") pod \"a772eddc-1462-4636-a4ba-1340b24dc102\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.336980 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-swift-storage-0\") pod \"a772eddc-1462-4636-a4ba-1340b24dc102\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.337071 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-svc\") pod \"a772eddc-1462-4636-a4ba-1340b24dc102\" (UID: \"a772eddc-1462-4636-a4ba-1340b24dc102\") " Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.362141 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a772eddc-1462-4636-a4ba-1340b24dc102-kube-api-access-d872c" (OuterVolumeSpecName: "kube-api-access-d872c") pod "a772eddc-1462-4636-a4ba-1340b24dc102" (UID: "a772eddc-1462-4636-a4ba-1340b24dc102"). InnerVolumeSpecName "kube-api-access-d872c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.402996 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a772eddc-1462-4636-a4ba-1340b24dc102" (UID: "a772eddc-1462-4636-a4ba-1340b24dc102"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.412548 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-config" (OuterVolumeSpecName: "config") pod "a772eddc-1462-4636-a4ba-1340b24dc102" (UID: "a772eddc-1462-4636-a4ba-1340b24dc102"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.418769 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a772eddc-1462-4636-a4ba-1340b24dc102" (UID: "a772eddc-1462-4636-a4ba-1340b24dc102"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.427741 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a772eddc-1462-4636-a4ba-1340b24dc102" (UID: "a772eddc-1462-4636-a4ba-1340b24dc102"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.438378 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.438411 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d872c\" (UniqueName: \"kubernetes.io/projected/a772eddc-1462-4636-a4ba-1340b24dc102-kube-api-access-d872c\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.438422 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.438431 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.438440 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.455581 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a772eddc-1462-4636-a4ba-1340b24dc102" (UID: "a772eddc-1462-4636-a4ba-1340b24dc102"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.513499 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.513579 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:15:58 crc kubenswrapper[4926]: I1007 21:15:58.540386 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a772eddc-1462-4636-a4ba-1340b24dc102-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.049950 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.050003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c5b587c5-jvms2" event={"ID":"a772eddc-1462-4636-a4ba-1340b24dc102","Type":"ContainerDied","Data":"572b636518e49533f8f9268c82d2d59424c52ca59da24ce9029fc7b6c2fcde86"} Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.050402 4926 scope.go:117] "RemoveContainer" containerID="ad0f1088b7fafad316bdcc107a5ed82e953a65ca275ebaf97ec7db01a4518852" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.082005 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84c5b587c5-jvms2"] Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.091912 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84c5b587c5-jvms2"] Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.096012 4926 scope.go:117] "RemoveContainer" containerID="e28592f2a9801ea3408b48216016b5e85e9c0138194b9419c3c0fa4eba009d6f" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.373547 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.373882 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.723286 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.863834 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-config-data\") pod \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.863897 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lq47\" (UniqueName: \"kubernetes.io/projected/a505eae2-6bc8-40cd-a133-9bd9bcb98478-kube-api-access-5lq47\") pod \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.863957 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-combined-ca-bundle\") pod \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.864031 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-scripts\") pod \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\" (UID: \"a505eae2-6bc8-40cd-a133-9bd9bcb98478\") " Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.870015 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a505eae2-6bc8-40cd-a133-9bd9bcb98478-kube-api-access-5lq47" (OuterVolumeSpecName: "kube-api-access-5lq47") pod "a505eae2-6bc8-40cd-a133-9bd9bcb98478" (UID: "a505eae2-6bc8-40cd-a133-9bd9bcb98478"). InnerVolumeSpecName "kube-api-access-5lq47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.871112 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-scripts" (OuterVolumeSpecName: "scripts") pod "a505eae2-6bc8-40cd-a133-9bd9bcb98478" (UID: "a505eae2-6bc8-40cd-a133-9bd9bcb98478"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.901212 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-config-data" (OuterVolumeSpecName: "config-data") pod "a505eae2-6bc8-40cd-a133-9bd9bcb98478" (UID: "a505eae2-6bc8-40cd-a133-9bd9bcb98478"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.909292 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a505eae2-6bc8-40cd-a133-9bd9bcb98478" (UID: "a505eae2-6bc8-40cd-a133-9bd9bcb98478"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.966291 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lq47\" (UniqueName: \"kubernetes.io/projected/a505eae2-6bc8-40cd-a133-9bd9bcb98478-kube-api-access-5lq47\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.966324 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.966339 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.966348 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a505eae2-6bc8-40cd-a133-9bd9bcb98478-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:15:59 crc kubenswrapper[4926]: I1007 21:15:59.980018 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.060136 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zl2hw" event={"ID":"a505eae2-6bc8-40cd-a133-9bd9bcb98478","Type":"ContainerDied","Data":"d103be10cbf04467dc1536d843697847311af7931411c94c6b43789dc5b860db"} Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.060174 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zl2hw" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.060183 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d103be10cbf04467dc1536d843697847311af7931411c94c6b43789dc5b860db" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.062104 4926 generic.go:334] "Generic (PLEG): container finished" podID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerID="eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5" exitCode=0 Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.062207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerDied","Data":"eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5"} Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.062218 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.062239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"392cbe25-f961-43ce-bc58-453dba0de3a9","Type":"ContainerDied","Data":"eee2d8930de42e94351b24102904da7250820c9feecac0c27babe13f169c075a"} Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.062271 4926 scope.go:117] "RemoveContainer" containerID="eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.067314 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-config-data\") pod \"392cbe25-f961-43ce-bc58-453dba0de3a9\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.067583 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-custom-prometheus-ca\") pod \"392cbe25-f961-43ce-bc58-453dba0de3a9\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.067708 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs4n8\" (UniqueName: \"kubernetes.io/projected/392cbe25-f961-43ce-bc58-453dba0de3a9-kube-api-access-fs4n8\") pod \"392cbe25-f961-43ce-bc58-453dba0de3a9\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.067835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/392cbe25-f961-43ce-bc58-453dba0de3a9-logs\") pod \"392cbe25-f961-43ce-bc58-453dba0de3a9\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.068012 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-combined-ca-bundle\") pod \"392cbe25-f961-43ce-bc58-453dba0de3a9\" (UID: \"392cbe25-f961-43ce-bc58-453dba0de3a9\") " Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.068133 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/392cbe25-f961-43ce-bc58-453dba0de3a9-logs" (OuterVolumeSpecName: "logs") pod "392cbe25-f961-43ce-bc58-453dba0de3a9" (UID: "392cbe25-f961-43ce-bc58-453dba0de3a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.069904 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/392cbe25-f961-43ce-bc58-453dba0de3a9-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.071063 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392cbe25-f961-43ce-bc58-453dba0de3a9-kube-api-access-fs4n8" (OuterVolumeSpecName: "kube-api-access-fs4n8") pod "392cbe25-f961-43ce-bc58-453dba0de3a9" (UID: "392cbe25-f961-43ce-bc58-453dba0de3a9"). InnerVolumeSpecName "kube-api-access-fs4n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.094360 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.118770 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "392cbe25-f961-43ce-bc58-453dba0de3a9" (UID: "392cbe25-f961-43ce-bc58-453dba0de3a9"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.175279 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "392cbe25-f961-43ce-bc58-453dba0de3a9" (UID: "392cbe25-f961-43ce-bc58-453dba0de3a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.175822 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.175845 4926 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.175853 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs4n8\" (UniqueName: \"kubernetes.io/projected/392cbe25-f961-43ce-bc58-453dba0de3a9-kube-api-access-fs4n8\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.202842 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.203110 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-log" containerID="cri-o://429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a" gracePeriod=30 Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.203225 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-api" containerID="cri-o://a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15" gracePeriod=30 Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.215881 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.216082 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="15217f6a-e2ad-4d8d-b675-e3aec437574f" containerName="nova-scheduler-scheduler" containerID="cri-o://7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25" gracePeriod=30 Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.232391 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-config-data" (OuterVolumeSpecName: "config-data") pod "392cbe25-f961-43ce-bc58-453dba0de3a9" (UID: "392cbe25-f961-43ce-bc58-453dba0de3a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.235295 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.277649 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/392cbe25-f961-43ce-bc58-453dba0de3a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.304407 4926 scope.go:117] "RemoveContainer" containerID="eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.304959 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5\": container with ID starting with eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5 not found: ID does not exist" containerID="eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.305006 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5"} err="failed to get container status \"eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5\": rpc error: code = NotFound desc = could not find container \"eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5\": container with ID starting with eef316b2aa062276efe53923afde854a3cb7ea13bdbe237777092c76b32593e5 not found: ID does not exist" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.305033 4926 scope.go:117] "RemoveContainer" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.305451 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb\": container with ID starting with d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb not found: ID does not exist" containerID="d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.305486 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb"} err="failed to get container status \"d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb\": rpc error: code = NotFound desc = could not find container \"d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb\": container with ID starting with d81712092191983b837176095b400b5e53256152a2228a4597310f4224fd21cb not found: ID does not exist" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.396535 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.414471 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427321 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.427786 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427807 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.427827 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a505eae2-6bc8-40cd-a133-9bd9bcb98478" containerName="nova-manage" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427836 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a505eae2-6bc8-40cd-a133-9bd9bcb98478" containerName="nova-manage" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.427851 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427860 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.427883 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" containerName="init" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427891 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" containerName="init" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.427912 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" containerName="dnsmasq-dns" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427920 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" containerName="dnsmasq-dns" Oct 07 21:16:00 crc kubenswrapper[4926]: E1007 21:16:00.427932 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.427950 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.428207 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.428229 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.428239 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.428253 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" containerName="dnsmasq-dns" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.428273 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a505eae2-6bc8-40cd-a133-9bd9bcb98478" containerName="nova-manage" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.429084 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.436930 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.448090 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.585306 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-logs\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.586880 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.587177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.587318 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr5bj\" (UniqueName: \"kubernetes.io/projected/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-kube-api-access-qr5bj\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.587419 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-config-data\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.695309 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-config-data\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.696283 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-logs\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.696442 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.696656 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.696746 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr5bj\" (UniqueName: \"kubernetes.io/projected/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-kube-api-access-qr5bj\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.697324 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-logs\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.697335 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" path="/var/lib/kubelet/pods/392cbe25-f961-43ce-bc58-453dba0de3a9/volumes" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.698373 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a772eddc-1462-4636-a4ba-1340b24dc102" path="/var/lib/kubelet/pods/a772eddc-1462-4636-a4ba-1340b24dc102/volumes" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.703780 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.704289 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-config-data\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.705766 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.719723 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr5bj\" (UniqueName: \"kubernetes.io/projected/d9756c29-8c2e-49c9-8bf2-f9a51710d7d7-kube-api-access-qr5bj\") pod \"watcher-decision-engine-0\" (UID: \"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7\") " pod="openstack/watcher-decision-engine-0" Oct 07 21:16:00 crc kubenswrapper[4926]: I1007 21:16:00.759334 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.081358 4926 generic.go:334] "Generic (PLEG): container finished" podID="111e3572-9e8a-4964-bc03-c5a50581e0f7" containerID="379872a104dbc4211471dfe06bd17267e6e5c5c3cbb195aaa20b1e77d143e971" exitCode=0 Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.081464 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbgql" event={"ID":"111e3572-9e8a-4964-bc03-c5a50581e0f7","Type":"ContainerDied","Data":"379872a104dbc4211471dfe06bd17267e6e5c5c3cbb195aaa20b1e77d143e971"} Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.083964 4926 generic.go:334] "Generic (PLEG): container finished" podID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerID="429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a" exitCode=143 Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.083997 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82","Type":"ContainerDied","Data":"429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a"} Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.084134 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-log" containerID="cri-o://af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4" gracePeriod=30 Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.084223 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-metadata" containerID="cri-o://c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916" gracePeriod=30 Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.212891 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.634955 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.827925 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-combined-ca-bundle\") pod \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.828135 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f2b397-8462-4fbf-a88f-40a4c803c3d0-logs\") pod \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.828160 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6wpc\" (UniqueName: \"kubernetes.io/projected/68f2b397-8462-4fbf-a88f-40a4c803c3d0-kube-api-access-f6wpc\") pod \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.828183 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-config-data\") pod \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.828233 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-nova-metadata-tls-certs\") pod \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\" (UID: \"68f2b397-8462-4fbf-a88f-40a4c803c3d0\") " Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.829621 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f2b397-8462-4fbf-a88f-40a4c803c3d0-logs" (OuterVolumeSpecName: "logs") pod "68f2b397-8462-4fbf-a88f-40a4c803c3d0" (UID: "68f2b397-8462-4fbf-a88f-40a4c803c3d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.834653 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f2b397-8462-4fbf-a88f-40a4c803c3d0-kube-api-access-f6wpc" (OuterVolumeSpecName: "kube-api-access-f6wpc") pod "68f2b397-8462-4fbf-a88f-40a4c803c3d0" (UID: "68f2b397-8462-4fbf-a88f-40a4c803c3d0"). InnerVolumeSpecName "kube-api-access-f6wpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.859523 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68f2b397-8462-4fbf-a88f-40a4c803c3d0" (UID: "68f2b397-8462-4fbf-a88f-40a4c803c3d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.871941 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-config-data" (OuterVolumeSpecName: "config-data") pod "68f2b397-8462-4fbf-a88f-40a4c803c3d0" (UID: "68f2b397-8462-4fbf-a88f-40a4c803c3d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.905890 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "68f2b397-8462-4fbf-a88f-40a4c803c3d0" (UID: "68f2b397-8462-4fbf-a88f-40a4c803c3d0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.930788 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f2b397-8462-4fbf-a88f-40a4c803c3d0-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.930834 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6wpc\" (UniqueName: \"kubernetes.io/projected/68f2b397-8462-4fbf-a88f-40a4c803c3d0-kube-api-access-f6wpc\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.930847 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.930856 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:01 crc kubenswrapper[4926]: I1007 21:16:01.930864 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2b397-8462-4fbf-a88f-40a4c803c3d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.097587 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7","Type":"ContainerStarted","Data":"2e92a90bdac55a7fc962af3c316f3d0e16fcd119246c6a20b27357ad24ed8747"} Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.097642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"d9756c29-8c2e-49c9-8bf2-f9a51710d7d7","Type":"ContainerStarted","Data":"46896790973d3732b3c219b3910420cb658901bda7ec53408e02cb0308b38422"} Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.099650 4926 generic.go:334] "Generic (PLEG): container finished" podID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerID="c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916" exitCode=0 Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.099674 4926 generic.go:334] "Generic (PLEG): container finished" podID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerID="af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4" exitCode=143 Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.099843 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.103304 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f2b397-8462-4fbf-a88f-40a4c803c3d0","Type":"ContainerDied","Data":"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916"} Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.103378 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f2b397-8462-4fbf-a88f-40a4c803c3d0","Type":"ContainerDied","Data":"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4"} Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.103394 4926 scope.go:117] "RemoveContainer" containerID="c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.103404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f2b397-8462-4fbf-a88f-40a4c803c3d0","Type":"ContainerDied","Data":"dc9c2aeee3bfbceae7eeb33effa556b0a94f8e97cbd251e2eb12ae93e943b4dc"} Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.135947 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.135924814 podStartE2EDuration="2.135924814s" podCreationTimestamp="2025-10-07 21:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:02.127606054 +0000 UTC m=+1212.166187244" watchObservedRunningTime="2025-10-07 21:16:02.135924814 +0000 UTC m=+1212.174505964" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.157150 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.171010 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.190314 4926 scope.go:117] "RemoveContainer" containerID="af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.190466 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.190954 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.190973 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.190995 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-log" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191005 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-log" Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.191029 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-metadata" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191037 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-metadata" Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.191073 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191084 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191368 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-metadata" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191405 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" containerName="nova-metadata-log" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191430 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.191927 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="392cbe25-f961-43ce-bc58-453dba0de3a9" containerName="watcher-decision-engine" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.192797 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.195596 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.195838 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.205923 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.225134 4926 scope.go:117] "RemoveContainer" containerID="c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916" Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.225696 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916\": container with ID starting with c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916 not found: ID does not exist" containerID="c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.225727 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916"} err="failed to get container status \"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916\": rpc error: code = NotFound desc = could not find container \"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916\": container with ID starting with c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916 not found: ID does not exist" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.226661 4926 scope.go:117] "RemoveContainer" containerID="af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4" Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.227038 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4\": container with ID starting with af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4 not found: ID does not exist" containerID="af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.227079 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4"} err="failed to get container status \"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4\": rpc error: code = NotFound desc = could not find container \"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4\": container with ID starting with af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4 not found: ID does not exist" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.227110 4926 scope.go:117] "RemoveContainer" containerID="c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.227470 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916"} err="failed to get container status \"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916\": rpc error: code = NotFound desc = could not find container \"c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916\": container with ID starting with c7d7fca8aa7ab507384a8e4fd581a725dee4252335ec40527e864950bbedb916 not found: ID does not exist" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.227520 4926 scope.go:117] "RemoveContainer" containerID="af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.227812 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4"} err="failed to get container status \"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4\": rpc error: code = NotFound desc = could not find container \"af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4\": container with ID starting with af8ebc7b6f394a8ccc56ebb1b0386a20b2b72cead9e0abef100f2f08589a2cf4 not found: ID does not exist" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.237968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-config-data\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.238062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35df0b0f-c822-4229-9c0b-f132d8ae09ee-logs\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.345993 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35df0b0f-c822-4229-9c0b-f132d8ae09ee-logs\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.346132 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6wwp\" (UniqueName: \"kubernetes.io/projected/35df0b0f-c822-4229-9c0b-f132d8ae09ee-kube-api-access-n6wwp\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.346161 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.346386 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-config-data\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.346451 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35df0b0f-c822-4229-9c0b-f132d8ae09ee-logs\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.346534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.357021 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-config-data\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.452040 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.452457 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6wwp\" (UniqueName: \"kubernetes.io/projected/35df0b0f-c822-4229-9c0b-f132d8ae09ee-kube-api-access-n6wwp\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.452482 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.465466 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.472122 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.476359 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6wwp\" (UniqueName: \"kubernetes.io/projected/35df0b0f-c822-4229-9c0b-f132d8ae09ee-kube-api-access-n6wwp\") pod \"nova-metadata-0\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.478111 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.479663 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.480855 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 21:16:02 crc kubenswrapper[4926]: E1007 21:16:02.480901 4926 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="15217f6a-e2ad-4d8d-b675-e3aec437574f" containerName="nova-scheduler-scheduler" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.509986 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.571970 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.655259 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-combined-ca-bundle\") pod \"111e3572-9e8a-4964-bc03-c5a50581e0f7\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.655439 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-scripts\") pod \"111e3572-9e8a-4964-bc03-c5a50581e0f7\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.655536 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-config-data\") pod \"111e3572-9e8a-4964-bc03-c5a50581e0f7\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.655566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtccv\" (UniqueName: \"kubernetes.io/projected/111e3572-9e8a-4964-bc03-c5a50581e0f7-kube-api-access-dtccv\") pod \"111e3572-9e8a-4964-bc03-c5a50581e0f7\" (UID: \"111e3572-9e8a-4964-bc03-c5a50581e0f7\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.670342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-scripts" (OuterVolumeSpecName: "scripts") pod "111e3572-9e8a-4964-bc03-c5a50581e0f7" (UID: "111e3572-9e8a-4964-bc03-c5a50581e0f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.670366 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111e3572-9e8a-4964-bc03-c5a50581e0f7-kube-api-access-dtccv" (OuterVolumeSpecName: "kube-api-access-dtccv") pod "111e3572-9e8a-4964-bc03-c5a50581e0f7" (UID: "111e3572-9e8a-4964-bc03-c5a50581e0f7"). InnerVolumeSpecName "kube-api-access-dtccv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.700431 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68f2b397-8462-4fbf-a88f-40a4c803c3d0" path="/var/lib/kubelet/pods/68f2b397-8462-4fbf-a88f-40a4c803c3d0/volumes" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.705721 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-config-data" (OuterVolumeSpecName: "config-data") pod "111e3572-9e8a-4964-bc03-c5a50581e0f7" (UID: "111e3572-9e8a-4964-bc03-c5a50581e0f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.714421 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "111e3572-9e8a-4964-bc03-c5a50581e0f7" (UID: "111e3572-9e8a-4964-bc03-c5a50581e0f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.716889 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756324 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-config-data\") pod \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756392 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-logs\") pod \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756411 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-combined-ca-bundle\") pod \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756563 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzt8r\" (UniqueName: \"kubernetes.io/projected/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-kube-api-access-fzt8r\") pod \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\" (UID: \"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82\") " Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756892 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756914 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756923 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/111e3572-9e8a-4964-bc03-c5a50581e0f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.756933 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtccv\" (UniqueName: \"kubernetes.io/projected/111e3572-9e8a-4964-bc03-c5a50581e0f7-kube-api-access-dtccv\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.757121 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-logs" (OuterVolumeSpecName: "logs") pod "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" (UID: "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.762512 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-kube-api-access-fzt8r" (OuterVolumeSpecName: "kube-api-access-fzt8r") pod "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" (UID: "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82"). InnerVolumeSpecName "kube-api-access-fzt8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.782464 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" (UID: "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.794232 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-config-data" (OuterVolumeSpecName: "config-data") pod "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" (UID: "aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.858324 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzt8r\" (UniqueName: \"kubernetes.io/projected/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-kube-api-access-fzt8r\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.858363 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.858375 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:02 crc kubenswrapper[4926]: I1007 21:16:02.858388 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.029661 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: W1007 21:16:03.032685 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35df0b0f_c822_4229_9c0b_f132d8ae09ee.slice/crio-5990f05ce881eb0cc2d0ac5b3af5f847b9a8c6d8073e407569bcd09ef815823a WatchSource:0}: Error finding container 5990f05ce881eb0cc2d0ac5b3af5f847b9a8c6d8073e407569bcd09ef815823a: Status 404 returned error can't find the container with id 5990f05ce881eb0cc2d0ac5b3af5f847b9a8c6d8073e407569bcd09ef815823a Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.118568 4926 generic.go:334] "Generic (PLEG): container finished" podID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerID="a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15" exitCode=0 Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.119167 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82","Type":"ContainerDied","Data":"a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15"} Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.119549 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82","Type":"ContainerDied","Data":"e1a76134948a8486a11f46b43aebf127d0785488d5ada2f3a402c5c7b098e578"} Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.119665 4926 scope.go:117] "RemoveContainer" containerID="a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.119769 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.139748 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbgql" event={"ID":"111e3572-9e8a-4964-bc03-c5a50581e0f7","Type":"ContainerDied","Data":"fdd76e171ec5d0ed86e8487828b7f1133eed6c61f4fd96d55e9bad4289982ca9"} Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.139809 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd76e171ec5d0ed86e8487828b7f1133eed6c61f4fd96d55e9bad4289982ca9" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.139766 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbgql" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.141916 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35df0b0f-c822-4229-9c0b-f132d8ae09ee","Type":"ContainerStarted","Data":"5990f05ce881eb0cc2d0ac5b3af5f847b9a8c6d8073e407569bcd09ef815823a"} Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.192866 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: E1007 21:16:03.193636 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-log" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.193653 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-log" Oct 07 21:16:03 crc kubenswrapper[4926]: E1007 21:16:03.193669 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-api" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.193700 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-api" Oct 07 21:16:03 crc kubenswrapper[4926]: E1007 21:16:03.193767 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111e3572-9e8a-4964-bc03-c5a50581e0f7" containerName="nova-cell1-conductor-db-sync" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.193777 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="111e3572-9e8a-4964-bc03-c5a50581e0f7" containerName="nova-cell1-conductor-db-sync" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.194156 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-log" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.194180 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="111e3572-9e8a-4964-bc03-c5a50581e0f7" containerName="nova-cell1-conductor-db-sync" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.194213 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" containerName="nova-api-api" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.195042 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.201160 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.205928 4926 scope.go:117] "RemoveContainer" containerID="429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.231089 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.247185 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.262176 4926 scope.go:117] "RemoveContainer" containerID="a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15" Oct 07 21:16:03 crc kubenswrapper[4926]: E1007 21:16:03.263721 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15\": container with ID starting with a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15 not found: ID does not exist" containerID="a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.263821 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15"} err="failed to get container status \"a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15\": rpc error: code = NotFound desc = could not find container \"a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15\": container with ID starting with a1474c5ec13417d4a59309a8bca30a70153e8471d05498b8ad20e3af93eefd15 not found: ID does not exist" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.263901 4926 scope.go:117] "RemoveContainer" containerID="429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a" Oct 07 21:16:03 crc kubenswrapper[4926]: E1007 21:16:03.264344 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a\": container with ID starting with 429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a not found: ID does not exist" containerID="429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.264449 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a"} err="failed to get container status \"429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a\": rpc error: code = NotFound desc = could not find container \"429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a\": container with ID starting with 429d26cf1b1ae4b605bcf7cdfa689f743ba56c6bf40d42424e0408345f40990a not found: ID does not exist" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.275288 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.299558 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.301174 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.302968 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.315914 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.370116 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6780f7-788b-42cf-937e-750a222de004-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.370697 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfm9t\" (UniqueName: \"kubernetes.io/projected/2f6780f7-788b-42cf-937e-750a222de004-kube-api-access-gfm9t\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.371085 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6780f7-788b-42cf-937e-750a222de004-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472570 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6780f7-788b-42cf-937e-750a222de004-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472618 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9rzw\" (UniqueName: \"kubernetes.io/projected/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-kube-api-access-h9rzw\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472633 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-config-data\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472656 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6780f7-788b-42cf-937e-750a222de004-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-logs\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472831 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfm9t\" (UniqueName: \"kubernetes.io/projected/2f6780f7-788b-42cf-937e-750a222de004-kube-api-access-gfm9t\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.472884 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.477813 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6780f7-788b-42cf-937e-750a222de004-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.477954 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6780f7-788b-42cf-937e-750a222de004-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.490451 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfm9t\" (UniqueName: \"kubernetes.io/projected/2f6780f7-788b-42cf-937e-750a222de004-kube-api-access-gfm9t\") pod \"nova-cell1-conductor-0\" (UID: \"2f6780f7-788b-42cf-937e-750a222de004\") " pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.519549 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.575825 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9rzw\" (UniqueName: \"kubernetes.io/projected/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-kube-api-access-h9rzw\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.575866 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-config-data\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.575902 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-logs\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.575990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.576642 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-logs\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.580971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.583733 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-config-data\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.591748 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9rzw\" (UniqueName: \"kubernetes.io/projected/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-kube-api-access-h9rzw\") pod \"nova-api-0\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " pod="openstack/nova-api-0" Oct 07 21:16:03 crc kubenswrapper[4926]: I1007 21:16:03.622621 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.004906 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.155975 4926 generic.go:334] "Generic (PLEG): container finished" podID="15217f6a-e2ad-4d8d-b675-e3aec437574f" containerID="7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25" exitCode=0 Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.156050 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15217f6a-e2ad-4d8d-b675-e3aec437574f","Type":"ContainerDied","Data":"7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25"} Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.156097 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15217f6a-e2ad-4d8d-b675-e3aec437574f","Type":"ContainerDied","Data":"7da87950545703d3308029067e58f7dbe43afdfce2f9fe2eb4a0602efade4e47"} Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.156109 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7da87950545703d3308029067e58f7dbe43afdfce2f9fe2eb4a0602efade4e47" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.159185 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35df0b0f-c822-4229-9c0b-f132d8ae09ee","Type":"ContainerStarted","Data":"d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516"} Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.159231 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35df0b0f-c822-4229-9c0b-f132d8ae09ee","Type":"ContainerStarted","Data":"87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a"} Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.169826 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2f6780f7-788b-42cf-937e-750a222de004","Type":"ContainerStarted","Data":"8a6691f5400e8ea88e3e2737fb2fb1bb65636d28284734d3299fb6e9d1f39a08"} Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.231360 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.269805 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.269773611 podStartE2EDuration="2.269773611s" podCreationTimestamp="2025-10-07 21:16:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:04.18341429 +0000 UTC m=+1214.221995440" watchObservedRunningTime="2025-10-07 21:16:04.269773611 +0000 UTC m=+1214.308354771" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.397575 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.400093 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh67s\" (UniqueName: \"kubernetes.io/projected/15217f6a-e2ad-4d8d-b675-e3aec437574f-kube-api-access-dh67s\") pod \"15217f6a-e2ad-4d8d-b675-e3aec437574f\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.400278 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-config-data\") pod \"15217f6a-e2ad-4d8d-b675-e3aec437574f\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.400420 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-combined-ca-bundle\") pod \"15217f6a-e2ad-4d8d-b675-e3aec437574f\" (UID: \"15217f6a-e2ad-4d8d-b675-e3aec437574f\") " Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.407546 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15217f6a-e2ad-4d8d-b675-e3aec437574f-kube-api-access-dh67s" (OuterVolumeSpecName: "kube-api-access-dh67s") pod "15217f6a-e2ad-4d8d-b675-e3aec437574f" (UID: "15217f6a-e2ad-4d8d-b675-e3aec437574f"). InnerVolumeSpecName "kube-api-access-dh67s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.427971 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-config-data" (OuterVolumeSpecName: "config-data") pod "15217f6a-e2ad-4d8d-b675-e3aec437574f" (UID: "15217f6a-e2ad-4d8d-b675-e3aec437574f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.439367 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15217f6a-e2ad-4d8d-b675-e3aec437574f" (UID: "15217f6a-e2ad-4d8d-b675-e3aec437574f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.501976 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.502004 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15217f6a-e2ad-4d8d-b675-e3aec437574f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.502015 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh67s\" (UniqueName: \"kubernetes.io/projected/15217f6a-e2ad-4d8d-b675-e3aec437574f-kube-api-access-dh67s\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:04 crc kubenswrapper[4926]: I1007 21:16:04.693634 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82" path="/var/lib/kubelet/pods/aaaf11b5-1eeb-4fd8-adc3-1c8e774e3f82/volumes" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.203125 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043","Type":"ContainerStarted","Data":"077117da8b6d82eccd31763b82d008a6476b1f9534acc31072167c90fffa82c2"} Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.204585 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043","Type":"ContainerStarted","Data":"5d7f43190c01557f3aa79392cd31a6468608f3aa9e36ade1cd77da3ef0fbbf3e"} Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.204690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043","Type":"ContainerStarted","Data":"2bcbe76d5ecf93c43fa66c3beaab6c0f36b5edad789c1ac5844128e79dfc6670"} Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.207445 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.207433 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2f6780f7-788b-42cf-937e-750a222de004","Type":"ContainerStarted","Data":"e7a1deeaeaf9049336a78b35026c26b346f668e1f457388faaf1bbaeb0f4889d"} Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.208756 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.225863 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.225842717 podStartE2EDuration="2.225842717s" podCreationTimestamp="2025-10-07 21:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:05.22178916 +0000 UTC m=+1215.260370320" watchObservedRunningTime="2025-10-07 21:16:05.225842717 +0000 UTC m=+1215.264423867" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.255105 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.268244 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.268381 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.268359573 podStartE2EDuration="2.268359573s" podCreationTimestamp="2025-10-07 21:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:05.253014901 +0000 UTC m=+1215.291596081" watchObservedRunningTime="2025-10-07 21:16:05.268359573 +0000 UTC m=+1215.306940733" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.299613 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:05 crc kubenswrapper[4926]: E1007 21:16:05.300379 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15217f6a-e2ad-4d8d-b675-e3aec437574f" containerName="nova-scheduler-scheduler" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.300506 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="15217f6a-e2ad-4d8d-b675-e3aec437574f" containerName="nova-scheduler-scheduler" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.300926 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="15217f6a-e2ad-4d8d-b675-e3aec437574f" containerName="nova-scheduler-scheduler" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.301869 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.305899 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.309369 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.419471 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.419749 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nbwd\" (UniqueName: \"kubernetes.io/projected/cd696bf0-2ccd-45b4-a498-698ad9406dc4-kube-api-access-4nbwd\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.419927 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-config-data\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.522318 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nbwd\" (UniqueName: \"kubernetes.io/projected/cd696bf0-2ccd-45b4-a498-698ad9406dc4-kube-api-access-4nbwd\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.522447 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-config-data\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.522851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.528666 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-config-data\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.532009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.544946 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nbwd\" (UniqueName: \"kubernetes.io/projected/cd696bf0-2ccd-45b4-a498-698ad9406dc4-kube-api-access-4nbwd\") pod \"nova-scheduler-0\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:05 crc kubenswrapper[4926]: I1007 21:16:05.620257 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:06 crc kubenswrapper[4926]: W1007 21:16:06.115394 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd696bf0_2ccd_45b4_a498_698ad9406dc4.slice/crio-80afd5b03259761ba992d4fc4d5f01ce2f9666020e5a3afbcf3ebf390429e04c WatchSource:0}: Error finding container 80afd5b03259761ba992d4fc4d5f01ce2f9666020e5a3afbcf3ebf390429e04c: Status 404 returned error can't find the container with id 80afd5b03259761ba992d4fc4d5f01ce2f9666020e5a3afbcf3ebf390429e04c Oct 07 21:16:06 crc kubenswrapper[4926]: I1007 21:16:06.130020 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:06 crc kubenswrapper[4926]: I1007 21:16:06.219222 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd696bf0-2ccd-45b4-a498-698ad9406dc4","Type":"ContainerStarted","Data":"80afd5b03259761ba992d4fc4d5f01ce2f9666020e5a3afbcf3ebf390429e04c"} Oct 07 21:16:06 crc kubenswrapper[4926]: I1007 21:16:06.700350 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15217f6a-e2ad-4d8d-b675-e3aec437574f" path="/var/lib/kubelet/pods/15217f6a-e2ad-4d8d-b675-e3aec437574f/volumes" Oct 07 21:16:07 crc kubenswrapper[4926]: I1007 21:16:07.230061 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd696bf0-2ccd-45b4-a498-698ad9406dc4","Type":"ContainerStarted","Data":"7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26"} Oct 07 21:16:07 crc kubenswrapper[4926]: I1007 21:16:07.253780 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.2537566079999998 podStartE2EDuration="2.253756608s" podCreationTimestamp="2025-10-07 21:16:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:07.25243446 +0000 UTC m=+1217.291015650" watchObservedRunningTime="2025-10-07 21:16:07.253756608 +0000 UTC m=+1217.292337768" Oct 07 21:16:07 crc kubenswrapper[4926]: I1007 21:16:07.510353 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 21:16:07 crc kubenswrapper[4926]: I1007 21:16:07.510577 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 21:16:07 crc kubenswrapper[4926]: I1007 21:16:07.878170 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 21:16:10 crc kubenswrapper[4926]: I1007 21:16:10.620701 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 21:16:10 crc kubenswrapper[4926]: I1007 21:16:10.760416 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:10 crc kubenswrapper[4926]: I1007 21:16:10.788506 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:11 crc kubenswrapper[4926]: I1007 21:16:11.275961 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:11 crc kubenswrapper[4926]: I1007 21:16:11.325234 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 07 21:16:11 crc kubenswrapper[4926]: I1007 21:16:11.808564 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:16:11 crc kubenswrapper[4926]: I1007 21:16:11.809071 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f838140a-347d-4bb3-8cc0-b4100bfe4b35" containerName="kube-state-metrics" containerID="cri-o://c5849ec77af20d02832aac47aa1f49f50eb8926ee7f12ca8b44110c99e92132d" gracePeriod=30 Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.288760 4926 generic.go:334] "Generic (PLEG): container finished" podID="f838140a-347d-4bb3-8cc0-b4100bfe4b35" containerID="c5849ec77af20d02832aac47aa1f49f50eb8926ee7f12ca8b44110c99e92132d" exitCode=2 Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.289750 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f838140a-347d-4bb3-8cc0-b4100bfe4b35","Type":"ContainerDied","Data":"c5849ec77af20d02832aac47aa1f49f50eb8926ee7f12ca8b44110c99e92132d"} Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.289859 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f838140a-347d-4bb3-8cc0-b4100bfe4b35","Type":"ContainerDied","Data":"5e5fc466a100ccc586d1c0fa2af9116b839ec50d83e10c27163504e1e4786994"} Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.289926 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e5fc466a100ccc586d1c0fa2af9116b839ec50d83e10c27163504e1e4786994" Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.346275 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.506386 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtk6r\" (UniqueName: \"kubernetes.io/projected/f838140a-347d-4bb3-8cc0-b4100bfe4b35-kube-api-access-wtk6r\") pod \"f838140a-347d-4bb3-8cc0-b4100bfe4b35\" (UID: \"f838140a-347d-4bb3-8cc0-b4100bfe4b35\") " Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.511242 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.511436 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.512350 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f838140a-347d-4bb3-8cc0-b4100bfe4b35-kube-api-access-wtk6r" (OuterVolumeSpecName: "kube-api-access-wtk6r") pod "f838140a-347d-4bb3-8cc0-b4100bfe4b35" (UID: "f838140a-347d-4bb3-8cc0-b4100bfe4b35"). InnerVolumeSpecName "kube-api-access-wtk6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:12 crc kubenswrapper[4926]: I1007 21:16:12.609388 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtk6r\" (UniqueName: \"kubernetes.io/projected/f838140a-347d-4bb3-8cc0-b4100bfe4b35-kube-api-access-wtk6r\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.295608 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.314333 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.323036 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.337797 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:16:13 crc kubenswrapper[4926]: E1007 21:16:13.338214 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f838140a-347d-4bb3-8cc0-b4100bfe4b35" containerName="kube-state-metrics" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.338228 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f838140a-347d-4bb3-8cc0-b4100bfe4b35" containerName="kube-state-metrics" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.338421 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f838140a-347d-4bb3-8cc0-b4100bfe4b35" containerName="kube-state-metrics" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.339070 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.340696 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.341219 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.351715 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.426696 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.427116 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.427148 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrtxb\" (UniqueName: \"kubernetes.io/projected/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-api-access-lrtxb\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.427184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.526694 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.526746 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.528401 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrtxb\" (UniqueName: \"kubernetes.io/projected/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-api-access-lrtxb\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.528449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.528558 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.528625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.533999 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.542976 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.558732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.562061 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.565362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrtxb\" (UniqueName: \"kubernetes.io/projected/b8c45bfd-abbf-432e-bb7f-70739b19c587-kube-api-access-lrtxb\") pod \"kube-state-metrics-0\" (UID: \"b8c45bfd-abbf-432e-bb7f-70739b19c587\") " pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.623427 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.623467 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.656921 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.785014 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.791168 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-central-agent" containerID="cri-o://658823b71079c86298e5a27d8a6eb88cd783ea6c2d3a65a0fbc934a4e52a44f3" gracePeriod=30 Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.791707 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="proxy-httpd" containerID="cri-o://8b4ddf88fe6eb7a6f2176d43fffd98a3c21e616d423be53c2f6066cc3b0a5465" gracePeriod=30 Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.791765 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="sg-core" containerID="cri-o://9f6d94f99d0ac2720efb728ceb1536625581eb30845ded5a9cb5483ecb6a4508" gracePeriod=30 Oct 07 21:16:13 crc kubenswrapper[4926]: I1007 21:16:13.791808 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-notification-agent" containerID="cri-o://8b34a8a9db2d6f0b0378f2fdb5c799275b840364fa246ed26d0c9c6740427741" gracePeriod=30 Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.227393 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.305619 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b8c45bfd-abbf-432e-bb7f-70739b19c587","Type":"ContainerStarted","Data":"08804d914c27e17d2221389dd7ed3267ddd5207db1099e7bc518c4681e40a245"} Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.308182 4926 generic.go:334] "Generic (PLEG): container finished" podID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerID="8b4ddf88fe6eb7a6f2176d43fffd98a3c21e616d423be53c2f6066cc3b0a5465" exitCode=0 Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.308320 4926 generic.go:334] "Generic (PLEG): container finished" podID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerID="9f6d94f99d0ac2720efb728ceb1536625581eb30845ded5a9cb5483ecb6a4508" exitCode=2 Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.308343 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerDied","Data":"8b4ddf88fe6eb7a6f2176d43fffd98a3c21e616d423be53c2f6066cc3b0a5465"} Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.308369 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerDied","Data":"9f6d94f99d0ac2720efb728ceb1536625581eb30845ded5a9cb5483ecb6a4508"} Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.706867 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.216:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.707323 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.216:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:14 crc kubenswrapper[4926]: I1007 21:16:14.739767 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f838140a-347d-4bb3-8cc0-b4100bfe4b35" path="/var/lib/kubelet/pods/f838140a-347d-4bb3-8cc0-b4100bfe4b35/volumes" Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.322694 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b8c45bfd-abbf-432e-bb7f-70739b19c587","Type":"ContainerStarted","Data":"d1f019afd74d13aea76ca92a404333449aa86ba9d3a763417745fbf219832bc9"} Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.323517 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.330045 4926 generic.go:334] "Generic (PLEG): container finished" podID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerID="658823b71079c86298e5a27d8a6eb88cd783ea6c2d3a65a0fbc934a4e52a44f3" exitCode=0 Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.330125 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerDied","Data":"658823b71079c86298e5a27d8a6eb88cd783ea6c2d3a65a0fbc934a4e52a44f3"} Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.354485 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.984044013 podStartE2EDuration="2.354466427s" podCreationTimestamp="2025-10-07 21:16:13 +0000 UTC" firstStartedPulling="2025-10-07 21:16:14.233288459 +0000 UTC m=+1224.271869609" lastFinishedPulling="2025-10-07 21:16:14.603710863 +0000 UTC m=+1224.642292023" observedRunningTime="2025-10-07 21:16:15.352449679 +0000 UTC m=+1225.391030839" watchObservedRunningTime="2025-10-07 21:16:15.354466427 +0000 UTC m=+1225.393047577" Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.621162 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 21:16:15 crc kubenswrapper[4926]: I1007 21:16:15.660975 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 21:16:16 crc kubenswrapper[4926]: I1007 21:16:16.374884 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.409956 4926 generic.go:334] "Generic (PLEG): container finished" podID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerID="8b34a8a9db2d6f0b0378f2fdb5c799275b840364fa246ed26d0c9c6740427741" exitCode=0 Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.410056 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerDied","Data":"8b34a8a9db2d6f0b0378f2fdb5c799275b840364fa246ed26d0c9c6740427741"} Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.555352 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.591851 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-sg-core-conf-yaml\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.591895 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-log-httpd\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.592360 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.592403 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-run-httpd\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.592455 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-config-data\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.592484 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-scripts\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.592715 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.593170 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.593219 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/000382f4-d2d7-4c63-aa0b-e14c64dca78c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.600333 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-scripts" (OuterVolumeSpecName: "scripts") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.658351 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.694219 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbj45\" (UniqueName: \"kubernetes.io/projected/000382f4-d2d7-4c63-aa0b-e14c64dca78c-kube-api-access-bbj45\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.694284 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-combined-ca-bundle\") pod \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\" (UID: \"000382f4-d2d7-4c63-aa0b-e14c64dca78c\") " Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.695622 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.695675 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.703896 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000382f4-d2d7-4c63-aa0b-e14c64dca78c-kube-api-access-bbj45" (OuterVolumeSpecName: "kube-api-access-bbj45") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "kube-api-access-bbj45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.744142 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-config-data" (OuterVolumeSpecName: "config-data") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.784981 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "000382f4-d2d7-4c63-aa0b-e14c64dca78c" (UID: "000382f4-d2d7-4c63-aa0b-e14c64dca78c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.797676 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.797828 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbj45\" (UniqueName: \"kubernetes.io/projected/000382f4-d2d7-4c63-aa0b-e14c64dca78c-kube-api-access-bbj45\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:20 crc kubenswrapper[4926]: I1007 21:16:20.797860 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000382f4-d2d7-4c63-aa0b-e14c64dca78c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.439459 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"000382f4-d2d7-4c63-aa0b-e14c64dca78c","Type":"ContainerDied","Data":"2b6a5fcecdb0fcef18200cdc386e0007f6c685ea584d72df7e34e611a4ca3e38"} Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.439539 4926 scope.go:117] "RemoveContainer" containerID="8b4ddf88fe6eb7a6f2176d43fffd98a3c21e616d423be53c2f6066cc3b0a5465" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.439558 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.478471 4926 scope.go:117] "RemoveContainer" containerID="9f6d94f99d0ac2720efb728ceb1536625581eb30845ded5a9cb5483ecb6a4508" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.511332 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.525256 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.535713 4926 scope.go:117] "RemoveContainer" containerID="8b34a8a9db2d6f0b0378f2fdb5c799275b840364fa246ed26d0c9c6740427741" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.562746 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:21 crc kubenswrapper[4926]: E1007 21:16:21.563322 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="sg-core" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563344 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="sg-core" Oct 07 21:16:21 crc kubenswrapper[4926]: E1007 21:16:21.563368 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-central-agent" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563377 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-central-agent" Oct 07 21:16:21 crc kubenswrapper[4926]: E1007 21:16:21.563391 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="proxy-httpd" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563399 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="proxy-httpd" Oct 07 21:16:21 crc kubenswrapper[4926]: E1007 21:16:21.563417 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-notification-agent" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563428 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-notification-agent" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563664 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="proxy-httpd" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563687 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-central-agent" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563715 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="ceilometer-notification-agent" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.563730 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" containerName="sg-core" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.566160 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.571556 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.571846 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.572005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.572641 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611094 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-config-data\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611136 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-scripts\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611160 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-run-httpd\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611217 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611252 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br55b\" (UniqueName: \"kubernetes.io/projected/0fe29a87-1276-464e-8a45-65b80f2c1fa6-kube-api-access-br55b\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611273 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-log-httpd\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611421 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.611618 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.612441 4926 scope.go:117] "RemoveContainer" containerID="658823b71079c86298e5a27d8a6eb88cd783ea6c2d3a65a0fbc934a4e52a44f3" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.712934 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713066 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-config-data\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-run-httpd\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713110 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-scripts\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br55b\" (UniqueName: \"kubernetes.io/projected/0fe29a87-1276-464e-8a45-65b80f2c1fa6-kube-api-access-br55b\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-log-httpd\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713244 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.713869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-run-httpd\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.714987 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-log-httpd\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.717719 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-scripts\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.719257 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.719375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.719999 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-config-data\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.720837 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.738099 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br55b\" (UniqueName: \"kubernetes.io/projected/0fe29a87-1276-464e-8a45-65b80f2c1fa6-kube-api-access-br55b\") pod \"ceilometer-0\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " pod="openstack/ceilometer-0" Oct 07 21:16:21 crc kubenswrapper[4926]: I1007 21:16:21.901577 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.417565 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.457381 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerStarted","Data":"d3c039f0451ff0238369679497ee33dc73d6f70f3dfe19a4320138c4413238ff"} Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.459415 4926 generic.go:334] "Generic (PLEG): container finished" podID="87abc81b-808f-4752-8ae9-5e34e609138b" containerID="4add980846fd6d10bc5891c0d57617d4f4c9ab3f7d342059614a29700e4e918d" exitCode=137 Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.459446 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"87abc81b-808f-4752-8ae9-5e34e609138b","Type":"ContainerDied","Data":"4add980846fd6d10bc5891c0d57617d4f4c9ab3f7d342059614a29700e4e918d"} Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.459461 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"87abc81b-808f-4752-8ae9-5e34e609138b","Type":"ContainerDied","Data":"7c15b9d613cbc1bebb488aa4622b7be2bf0ab5939c1eaed3079c8d5cf04f9c20"} Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.459473 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c15b9d613cbc1bebb488aa4622b7be2bf0ab5939c1eaed3079c8d5cf04f9c20" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.517569 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.518886 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.528822 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.630640 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.720468 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000382f4-d2d7-4c63-aa0b-e14c64dca78c" path="/var/lib/kubelet/pods/000382f4-d2d7-4c63-aa0b-e14c64dca78c/volumes" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.746210 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-combined-ca-bundle\") pod \"87abc81b-808f-4752-8ae9-5e34e609138b\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.746296 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-config-data\") pod \"87abc81b-808f-4752-8ae9-5e34e609138b\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.746346 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjlz2\" (UniqueName: \"kubernetes.io/projected/87abc81b-808f-4752-8ae9-5e34e609138b-kube-api-access-kjlz2\") pod \"87abc81b-808f-4752-8ae9-5e34e609138b\" (UID: \"87abc81b-808f-4752-8ae9-5e34e609138b\") " Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.753003 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87abc81b-808f-4752-8ae9-5e34e609138b-kube-api-access-kjlz2" (OuterVolumeSpecName: "kube-api-access-kjlz2") pod "87abc81b-808f-4752-8ae9-5e34e609138b" (UID: "87abc81b-808f-4752-8ae9-5e34e609138b"). InnerVolumeSpecName "kube-api-access-kjlz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.775116 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87abc81b-808f-4752-8ae9-5e34e609138b" (UID: "87abc81b-808f-4752-8ae9-5e34e609138b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.789163 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-config-data" (OuterVolumeSpecName: "config-data") pod "87abc81b-808f-4752-8ae9-5e34e609138b" (UID: "87abc81b-808f-4752-8ae9-5e34e609138b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.850425 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.850488 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87abc81b-808f-4752-8ae9-5e34e609138b-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:22 crc kubenswrapper[4926]: I1007 21:16:22.850500 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjlz2\" (UniqueName: \"kubernetes.io/projected/87abc81b-808f-4752-8ae9-5e34e609138b-kube-api-access-kjlz2\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.475689 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerStarted","Data":"90bcdb612080bafcc0512ac0328f157cecb3521975ac4bf5f6f43e66dac06b74"} Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.476489 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerStarted","Data":"e290f57455664b4537989136d45a58a29ce0ced33a6b58d4ac5560945a494c06"} Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.475742 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.501789 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.543093 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.570865 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.591880 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:16:23 crc kubenswrapper[4926]: E1007 21:16:23.592587 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87abc81b-808f-4752-8ae9-5e34e609138b" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.592611 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="87abc81b-808f-4752-8ae9-5e34e609138b" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.592868 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="87abc81b-808f-4752-8ae9-5e34e609138b" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.595081 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.597087 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.598377 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.600255 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.630366 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.633068 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.639159 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.640560 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.649375 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.689223 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.768937 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.769525 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.769554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.769707 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.769726 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvqlx\" (UniqueName: \"kubernetes.io/projected/6e980975-58c0-4255-bc3e-df92773323cf-kube-api-access-dvqlx\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.870832 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.870866 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvqlx\" (UniqueName: \"kubernetes.io/projected/6e980975-58c0-4255-bc3e-df92773323cf-kube-api-access-dvqlx\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.870920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.870950 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.870972 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.876046 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.879061 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.878635 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.879712 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e980975-58c0-4255-bc3e-df92773323cf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.895160 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvqlx\" (UniqueName: \"kubernetes.io/projected/6e980975-58c0-4255-bc3e-df92773323cf-kube-api-access-dvqlx\") pod \"nova-cell1-novncproxy-0\" (UID: \"6e980975-58c0-4255-bc3e-df92773323cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:23 crc kubenswrapper[4926]: I1007 21:16:23.993210 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.426916 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 21:16:24 crc kubenswrapper[4926]: W1007 21:16:24.436996 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e980975_58c0_4255_bc3e_df92773323cf.slice/crio-5e3152610eddc153c4e69d53c8288cdd2f7225208eae7ad2642c3e1af8b2baaa WatchSource:0}: Error finding container 5e3152610eddc153c4e69d53c8288cdd2f7225208eae7ad2642c3e1af8b2baaa: Status 404 returned error can't find the container with id 5e3152610eddc153c4e69d53c8288cdd2f7225208eae7ad2642c3e1af8b2baaa Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.490926 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerStarted","Data":"add9ddd9ac60f1f83b34f0d9dfc2a07470ac070131cf7422af083bbb8a465667"} Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.499572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6e980975-58c0-4255-bc3e-df92773323cf","Type":"ContainerStarted","Data":"5e3152610eddc153c4e69d53c8288cdd2f7225208eae7ad2642c3e1af8b2baaa"} Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.500241 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.510015 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.698173 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87abc81b-808f-4752-8ae9-5e34e609138b" path="/var/lib/kubelet/pods/87abc81b-808f-4752-8ae9-5e34e609138b/volumes" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.709251 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f9cb5965-h6hqc"] Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.749377 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f9cb5965-h6hqc"] Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.749484 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.893488 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.893543 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-svc\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.893573 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.893662 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-config\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.893692 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpzg5\" (UniqueName: \"kubernetes.io/projected/3079e3d9-da18-49f6-b29b-275d42b4885c-kube-api-access-mpzg5\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.893729 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.995542 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.996423 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-svc\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.996362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.996499 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.997144 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.997267 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-config\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.997305 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpzg5\" (UniqueName: \"kubernetes.io/projected/3079e3d9-da18-49f6-b29b-275d42b4885c-kube-api-access-mpzg5\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.997339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.997532 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-svc\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.997894 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:24 crc kubenswrapper[4926]: I1007 21:16:24.998088 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-config\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.015465 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpzg5\" (UniqueName: \"kubernetes.io/projected/3079e3d9-da18-49f6-b29b-275d42b4885c-kube-api-access-mpzg5\") pod \"dnsmasq-dns-85f9cb5965-h6hqc\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.082995 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.519259 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerStarted","Data":"f2e0b343ba4ba07a2df15f78eefd536a98b54a9f00dfdfd262c10e9161612710"} Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.519549 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.522444 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6e980975-58c0-4255-bc3e-df92773323cf","Type":"ContainerStarted","Data":"199a77b1fa3a5fad170f325adf024347b33743ade106ef6abb50520e822758d2"} Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.544402 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9159125160000001 podStartE2EDuration="4.544372599s" podCreationTimestamp="2025-10-07 21:16:21 +0000 UTC" firstStartedPulling="2025-10-07 21:16:22.436274217 +0000 UTC m=+1232.474855367" lastFinishedPulling="2025-10-07 21:16:25.0647343 +0000 UTC m=+1235.103315450" observedRunningTime="2025-10-07 21:16:25.542143474 +0000 UTC m=+1235.580724644" watchObservedRunningTime="2025-10-07 21:16:25.544372599 +0000 UTC m=+1235.582953739" Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.576659 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.576628726 podStartE2EDuration="2.576628726s" podCreationTimestamp="2025-10-07 21:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:25.564320119 +0000 UTC m=+1235.602901299" watchObservedRunningTime="2025-10-07 21:16:25.576628726 +0000 UTC m=+1235.615209886" Oct 07 21:16:25 crc kubenswrapper[4926]: I1007 21:16:25.599828 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f9cb5965-h6hqc"] Oct 07 21:16:26 crc kubenswrapper[4926]: I1007 21:16:26.532286 4926 generic.go:334] "Generic (PLEG): container finished" podID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerID="bbadce471ff9b97707c0094473fe87c379298d36580e666e73727c0d03562b80" exitCode=0 Oct 07 21:16:26 crc kubenswrapper[4926]: I1007 21:16:26.532390 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" event={"ID":"3079e3d9-da18-49f6-b29b-275d42b4885c","Type":"ContainerDied","Data":"bbadce471ff9b97707c0094473fe87c379298d36580e666e73727c0d03562b80"} Oct 07 21:16:26 crc kubenswrapper[4926]: I1007 21:16:26.532836 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" event={"ID":"3079e3d9-da18-49f6-b29b-275d42b4885c","Type":"ContainerStarted","Data":"10d2102e4b71bc3230485a81b62873dd734db2ee7b910542c0f4d265d0393f81"} Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.513410 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.527107 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.562043 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" event={"ID":"3079e3d9-da18-49f6-b29b-275d42b4885c","Type":"ContainerStarted","Data":"87c7b36a97a6de76ce763e116a6f3dab812136a114e90a2387da5730a3c6283e"} Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.562270 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-central-agent" containerID="cri-o://e290f57455664b4537989136d45a58a29ce0ced33a6b58d4ac5560945a494c06" gracePeriod=30 Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.562805 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="sg-core" containerID="cri-o://add9ddd9ac60f1f83b34f0d9dfc2a07470ac070131cf7422af083bbb8a465667" gracePeriod=30 Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.562863 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="proxy-httpd" containerID="cri-o://f2e0b343ba4ba07a2df15f78eefd536a98b54a9f00dfdfd262c10e9161612710" gracePeriod=30 Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.562901 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-notification-agent" containerID="cri-o://90bcdb612080bafcc0512ac0328f157cecb3521975ac4bf5f6f43e66dac06b74" gracePeriod=30 Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.563177 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-log" containerID="cri-o://5d7f43190c01557f3aa79392cd31a6468608f3aa9e36ade1cd77da3ef0fbbf3e" gracePeriod=30 Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.563276 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-api" containerID="cri-o://077117da8b6d82eccd31763b82d008a6476b1f9534acc31072167c90fffa82c2" gracePeriod=30 Oct 07 21:16:27 crc kubenswrapper[4926]: I1007 21:16:27.595036 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" podStartSLOduration=3.59501777 podStartE2EDuration="3.59501777s" podCreationTimestamp="2025-10-07 21:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:27.589095168 +0000 UTC m=+1237.627676318" watchObservedRunningTime="2025-10-07 21:16:27.59501777 +0000 UTC m=+1237.633598920" Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.573989 4926 generic.go:334] "Generic (PLEG): container finished" podID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerID="f2e0b343ba4ba07a2df15f78eefd536a98b54a9f00dfdfd262c10e9161612710" exitCode=0 Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.574335 4926 generic.go:334] "Generic (PLEG): container finished" podID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerID="add9ddd9ac60f1f83b34f0d9dfc2a07470ac070131cf7422af083bbb8a465667" exitCode=2 Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.574349 4926 generic.go:334] "Generic (PLEG): container finished" podID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerID="90bcdb612080bafcc0512ac0328f157cecb3521975ac4bf5f6f43e66dac06b74" exitCode=0 Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.574048 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerDied","Data":"f2e0b343ba4ba07a2df15f78eefd536a98b54a9f00dfdfd262c10e9161612710"} Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.574391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerDied","Data":"add9ddd9ac60f1f83b34f0d9dfc2a07470ac070131cf7422af083bbb8a465667"} Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.574406 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerDied","Data":"90bcdb612080bafcc0512ac0328f157cecb3521975ac4bf5f6f43e66dac06b74"} Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.576753 4926 generic.go:334] "Generic (PLEG): container finished" podID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerID="5d7f43190c01557f3aa79392cd31a6468608f3aa9e36ade1cd77da3ef0fbbf3e" exitCode=143 Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.576896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043","Type":"ContainerDied","Data":"5d7f43190c01557f3aa79392cd31a6468608f3aa9e36ade1cd77da3ef0fbbf3e"} Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.577035 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:28 crc kubenswrapper[4926]: I1007 21:16:28.994382 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.590765 4926 generic.go:334] "Generic (PLEG): container finished" podID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerID="077117da8b6d82eccd31763b82d008a6476b1f9534acc31072167c90fffa82c2" exitCode=0 Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.590944 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043","Type":"ContainerDied","Data":"077117da8b6d82eccd31763b82d008a6476b1f9534acc31072167c90fffa82c2"} Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.592004 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043","Type":"ContainerDied","Data":"2bcbe76d5ecf93c43fa66c3beaab6c0f36b5edad789c1ac5844128e79dfc6670"} Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.592020 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bcbe76d5ecf93c43fa66c3beaab6c0f36b5edad789c1ac5844128e79dfc6670" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.622449 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.709580 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-combined-ca-bundle\") pod \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.709653 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-logs\") pod \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.709824 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-config-data\") pod \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.709859 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9rzw\" (UniqueName: \"kubernetes.io/projected/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-kube-api-access-h9rzw\") pod \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\" (UID: \"8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043\") " Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.710255 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-logs" (OuterVolumeSpecName: "logs") pod "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" (UID: "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.710713 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.720715 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-kube-api-access-h9rzw" (OuterVolumeSpecName: "kube-api-access-h9rzw") pod "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" (UID: "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043"). InnerVolumeSpecName "kube-api-access-h9rzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.743833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" (UID: "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.749667 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-config-data" (OuterVolumeSpecName: "config-data") pod "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" (UID: "8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.812391 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.812417 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9rzw\" (UniqueName: \"kubernetes.io/projected/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-kube-api-access-h9rzw\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:29 crc kubenswrapper[4926]: I1007 21:16:29.812429 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.602759 4926 generic.go:334] "Generic (PLEG): container finished" podID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerID="e290f57455664b4537989136d45a58a29ce0ced33a6b58d4ac5560945a494c06" exitCode=0 Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.602818 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerDied","Data":"e290f57455664b4537989136d45a58a29ce0ced33a6b58d4ac5560945a494c06"} Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.603105 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.650594 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.668307 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.673893 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:30 crc kubenswrapper[4926]: E1007 21:16:30.674481 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-log" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.674507 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-log" Oct 07 21:16:30 crc kubenswrapper[4926]: E1007 21:16:30.674556 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-api" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.674565 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-api" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.674799 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-api" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.674831 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" containerName="nova-api-log" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.676937 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.683100 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.683362 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.686492 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.709770 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043" path="/var/lib/kubelet/pods/8c13f7e2-d7de-44a1-ba4e-1e3dc8b7a043/volumes" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.710607 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.838623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000f063a-c62d-48dc-92a0-b5b6f317d60b-logs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.838668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-config-data\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.838787 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.838972 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.839005 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsm4l\" (UniqueName: \"kubernetes.io/projected/000f063a-c62d-48dc-92a0-b5b6f317d60b-kube-api-access-jsm4l\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.839184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-public-tls-certs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.865168 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.941333 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000f063a-c62d-48dc-92a0-b5b6f317d60b-logs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.941377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-config-data\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.941397 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.941442 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.941464 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsm4l\" (UniqueName: \"kubernetes.io/projected/000f063a-c62d-48dc-92a0-b5b6f317d60b-kube-api-access-jsm4l\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.941507 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-public-tls-certs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.942241 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000f063a-c62d-48dc-92a0-b5b6f317d60b-logs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.947555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.949327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-public-tls-certs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.953566 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-config-data\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.953905 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:30 crc kubenswrapper[4926]: I1007 21:16:30.961592 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsm4l\" (UniqueName: \"kubernetes.io/projected/000f063a-c62d-48dc-92a0-b5b6f317d60b-kube-api-access-jsm4l\") pod \"nova-api-0\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " pod="openstack/nova-api-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.018981 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.042904 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-config-data\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043007 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-log-httpd\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043027 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-sg-core-conf-yaml\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043058 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-run-httpd\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-ceilometer-tls-certs\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043146 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-scripts\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043169 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br55b\" (UniqueName: \"kubernetes.io/projected/0fe29a87-1276-464e-8a45-65b80f2c1fa6-kube-api-access-br55b\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043240 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-combined-ca-bundle\") pod \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\" (UID: \"0fe29a87-1276-464e-8a45-65b80f2c1fa6\") " Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043578 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.043853 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.044756 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.044774 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fe29a87-1276-464e-8a45-65b80f2c1fa6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.046823 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fe29a87-1276-464e-8a45-65b80f2c1fa6-kube-api-access-br55b" (OuterVolumeSpecName: "kube-api-access-br55b") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "kube-api-access-br55b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.047237 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-scripts" (OuterVolumeSpecName: "scripts") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.073777 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.132162 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.147657 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.147687 4926 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.147699 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.147709 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br55b\" (UniqueName: \"kubernetes.io/projected/0fe29a87-1276-464e-8a45-65b80f2c1fa6-kube-api-access-br55b\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.163277 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.240171 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-config-data" (OuterVolumeSpecName: "config-data") pod "0fe29a87-1276-464e-8a45-65b80f2c1fa6" (UID: "0fe29a87-1276-464e-8a45-65b80f2c1fa6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.249465 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.249494 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fe29a87-1276-464e-8a45-65b80f2c1fa6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.513844 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.617433 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fe29a87-1276-464e-8a45-65b80f2c1fa6","Type":"ContainerDied","Data":"d3c039f0451ff0238369679497ee33dc73d6f70f3dfe19a4320138c4413238ff"} Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.617510 4926 scope.go:117] "RemoveContainer" containerID="f2e0b343ba4ba07a2df15f78eefd536a98b54a9f00dfdfd262c10e9161612710" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.617729 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.623904 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"000f063a-c62d-48dc-92a0-b5b6f317d60b","Type":"ContainerStarted","Data":"d4476f5b51e499504309d1eb4ba56059882eb0d3bca36bc23ee62de924486eb7"} Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.670179 4926 scope.go:117] "RemoveContainer" containerID="add9ddd9ac60f1f83b34f0d9dfc2a07470ac070131cf7422af083bbb8a465667" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.701751 4926 scope.go:117] "RemoveContainer" containerID="90bcdb612080bafcc0512ac0328f157cecb3521975ac4bf5f6f43e66dac06b74" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.706556 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.719953 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.726098 4926 scope.go:117] "RemoveContainer" containerID="e290f57455664b4537989136d45a58a29ce0ced33a6b58d4ac5560945a494c06" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.732236 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:31 crc kubenswrapper[4926]: E1007 21:16:31.732708 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-notification-agent" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.732726 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-notification-agent" Oct 07 21:16:31 crc kubenswrapper[4926]: E1007 21:16:31.732748 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-central-agent" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.732755 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-central-agent" Oct 07 21:16:31 crc kubenswrapper[4926]: E1007 21:16:31.732776 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="proxy-httpd" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.732783 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="proxy-httpd" Oct 07 21:16:31 crc kubenswrapper[4926]: E1007 21:16:31.732807 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="sg-core" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.732813 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="sg-core" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.733002 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-central-agent" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.733019 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="proxy-httpd" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.733029 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="ceilometer-notification-agent" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.733036 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" containerName="sg-core" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.734937 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.738960 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.739827 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.740183 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.740390 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.858740 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-scripts\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.858815 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69db6fef-77ff-4e55-9a3c-2df537ae1632-run-httpd\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.858982 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-config-data\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.859016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69db6fef-77ff-4e55-9a3c-2df537ae1632-log-httpd\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.859059 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5jj6\" (UniqueName: \"kubernetes.io/projected/69db6fef-77ff-4e55-9a3c-2df537ae1632-kube-api-access-x5jj6\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.859101 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.859177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.859335 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960819 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960880 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-scripts\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69db6fef-77ff-4e55-9a3c-2df537ae1632-run-httpd\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960943 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-config-data\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960957 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69db6fef-77ff-4e55-9a3c-2df537ae1632-log-httpd\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960978 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5jj6\" (UniqueName: \"kubernetes.io/projected/69db6fef-77ff-4e55-9a3c-2df537ae1632-kube-api-access-x5jj6\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.960998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.961036 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.961371 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69db6fef-77ff-4e55-9a3c-2df537ae1632-log-httpd\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.962546 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69db6fef-77ff-4e55-9a3c-2df537ae1632-run-httpd\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.965861 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.965883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.966804 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-scripts\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.967404 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.968690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69db6fef-77ff-4e55-9a3c-2df537ae1632-config-data\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:31 crc kubenswrapper[4926]: I1007 21:16:31.978558 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5jj6\" (UniqueName: \"kubernetes.io/projected/69db6fef-77ff-4e55-9a3c-2df537ae1632-kube-api-access-x5jj6\") pod \"ceilometer-0\" (UID: \"69db6fef-77ff-4e55-9a3c-2df537ae1632\") " pod="openstack/ceilometer-0" Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.057125 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.546283 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 21:16:32 crc kubenswrapper[4926]: W1007 21:16:32.547669 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69db6fef_77ff_4e55_9a3c_2df537ae1632.slice/crio-3d037af1a1b1828a133ae5a8b062c5c6b158d936f31a0a60df8209cd979d8f11 WatchSource:0}: Error finding container 3d037af1a1b1828a133ae5a8b062c5c6b158d936f31a0a60df8209cd979d8f11: Status 404 returned error can't find the container with id 3d037af1a1b1828a133ae5a8b062c5c6b158d936f31a0a60df8209cd979d8f11 Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.638294 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"000f063a-c62d-48dc-92a0-b5b6f317d60b","Type":"ContainerStarted","Data":"a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd"} Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.638735 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"000f063a-c62d-48dc-92a0-b5b6f317d60b","Type":"ContainerStarted","Data":"c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f"} Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.639862 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69db6fef-77ff-4e55-9a3c-2df537ae1632","Type":"ContainerStarted","Data":"3d037af1a1b1828a133ae5a8b062c5c6b158d936f31a0a60df8209cd979d8f11"} Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.669963 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.669934787 podStartE2EDuration="2.669934787s" podCreationTimestamp="2025-10-07 21:16:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:32.659403781 +0000 UTC m=+1242.697984931" watchObservedRunningTime="2025-10-07 21:16:32.669934787 +0000 UTC m=+1242.708515977" Oct 07 21:16:32 crc kubenswrapper[4926]: I1007 21:16:32.694158 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fe29a87-1276-464e-8a45-65b80f2c1fa6" path="/var/lib/kubelet/pods/0fe29a87-1276-464e-8a45-65b80f2c1fa6/volumes" Oct 07 21:16:33 crc kubenswrapper[4926]: I1007 21:16:33.650353 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69db6fef-77ff-4e55-9a3c-2df537ae1632","Type":"ContainerStarted","Data":"5949ac7ba0082184c3be0c7448c1ea20456671387a994fa8bad6840ca7105196"} Oct 07 21:16:33 crc kubenswrapper[4926]: I1007 21:16:33.650690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69db6fef-77ff-4e55-9a3c-2df537ae1632","Type":"ContainerStarted","Data":"e9445c151b2233dd4e0274f90223a8c3c7405d12c42dcaa77ed20181c2061b88"} Oct 07 21:16:33 crc kubenswrapper[4926]: I1007 21:16:33.994025 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.018313 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.666907 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69db6fef-77ff-4e55-9a3c-2df537ae1632","Type":"ContainerStarted","Data":"345bf8405deb3324ccd559d80168752bdd157c56c7d9643c707cefc2fca8db8e"} Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.697915 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.930656 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-l8s8j"] Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.932872 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.936290 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.940035 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 07 21:16:34 crc kubenswrapper[4926]: I1007 21:16:34.951102 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-l8s8j"] Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.027361 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ksbx\" (UniqueName: \"kubernetes.io/projected/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-kube-api-access-9ksbx\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.027658 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-config-data\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.027903 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.028032 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-scripts\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.084484 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.129827 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.129883 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-scripts\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.129953 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ksbx\" (UniqueName: \"kubernetes.io/projected/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-kube-api-access-9ksbx\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.130038 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-config-data\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.137884 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-scripts\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.141411 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-config-data\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.156789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.165733 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ksbx\" (UniqueName: \"kubernetes.io/projected/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-kube-api-access-9ksbx\") pod \"nova-cell1-cell-mapping-l8s8j\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.183327 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8685f8ddff-w54pw"] Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.186827 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" containerName="dnsmasq-dns" containerID="cri-o://a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da" gracePeriod=10 Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.256068 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.717636 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.723283 4926 generic.go:334] "Generic (PLEG): container finished" podID="3cf26616-4656-4a96-b782-fb17eb11b912" containerID="a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da" exitCode=0 Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.723364 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" event={"ID":"3cf26616-4656-4a96-b782-fb17eb11b912","Type":"ContainerDied","Data":"a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da"} Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.723392 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" event={"ID":"3cf26616-4656-4a96-b782-fb17eb11b912","Type":"ContainerDied","Data":"7865f89c3c3c2175990d1a7d3dc742f53b1719988ea18c17237475c6185b65f2"} Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.723409 4926 scope.go:117] "RemoveContainer" containerID="a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.744735 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69db6fef-77ff-4e55-9a3c-2df537ae1632","Type":"ContainerStarted","Data":"8519cf77719f3d211e1d51bd49c41b9da45ca1123de4d2757210436b0ca9d91d"} Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.744825 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.766101 4926 scope.go:117] "RemoveContainer" containerID="cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.769774 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.093044113 podStartE2EDuration="4.769758018s" podCreationTimestamp="2025-10-07 21:16:31 +0000 UTC" firstStartedPulling="2025-10-07 21:16:32.551754213 +0000 UTC m=+1242.590335403" lastFinishedPulling="2025-10-07 21:16:35.228468158 +0000 UTC m=+1245.267049308" observedRunningTime="2025-10-07 21:16:35.768938525 +0000 UTC m=+1245.807519675" watchObservedRunningTime="2025-10-07 21:16:35.769758018 +0000 UTC m=+1245.808339178" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.795415 4926 scope.go:117] "RemoveContainer" containerID="a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da" Oct 07 21:16:35 crc kubenswrapper[4926]: E1007 21:16:35.795764 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da\": container with ID starting with a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da not found: ID does not exist" containerID="a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.795795 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da"} err="failed to get container status \"a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da\": rpc error: code = NotFound desc = could not find container \"a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da\": container with ID starting with a8ab2bb38d820fee6916edc254badbb1082da8d1f82fa5ef99202afd9340b3da not found: ID does not exist" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.795817 4926 scope.go:117] "RemoveContainer" containerID="cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8" Oct 07 21:16:35 crc kubenswrapper[4926]: E1007 21:16:35.795984 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8\": container with ID starting with cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8 not found: ID does not exist" containerID="cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.796005 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8"} err="failed to get container status \"cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8\": rpc error: code = NotFound desc = could not find container \"cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8\": container with ID starting with cffa7f7c87327f8553af3170ca946e05f7f2b06e61edbe65b80c411303fa47c8 not found: ID does not exist" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.848732 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-nb\") pod \"3cf26616-4656-4a96-b782-fb17eb11b912\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.848783 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltm58\" (UniqueName: \"kubernetes.io/projected/3cf26616-4656-4a96-b782-fb17eb11b912-kube-api-access-ltm58\") pod \"3cf26616-4656-4a96-b782-fb17eb11b912\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.848824 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-sb\") pod \"3cf26616-4656-4a96-b782-fb17eb11b912\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.848890 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-swift-storage-0\") pod \"3cf26616-4656-4a96-b782-fb17eb11b912\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.848921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-svc\") pod \"3cf26616-4656-4a96-b782-fb17eb11b912\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.849001 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-config\") pod \"3cf26616-4656-4a96-b782-fb17eb11b912\" (UID: \"3cf26616-4656-4a96-b782-fb17eb11b912\") " Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.855222 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf26616-4656-4a96-b782-fb17eb11b912-kube-api-access-ltm58" (OuterVolumeSpecName: "kube-api-access-ltm58") pod "3cf26616-4656-4a96-b782-fb17eb11b912" (UID: "3cf26616-4656-4a96-b782-fb17eb11b912"). InnerVolumeSpecName "kube-api-access-ltm58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.875574 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-l8s8j"] Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.945278 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3cf26616-4656-4a96-b782-fb17eb11b912" (UID: "3cf26616-4656-4a96-b782-fb17eb11b912"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.949818 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-config" (OuterVolumeSpecName: "config") pod "3cf26616-4656-4a96-b782-fb17eb11b912" (UID: "3cf26616-4656-4a96-b782-fb17eb11b912"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.952563 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.952594 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.952607 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltm58\" (UniqueName: \"kubernetes.io/projected/3cf26616-4656-4a96-b782-fb17eb11b912-kube-api-access-ltm58\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.952974 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3cf26616-4656-4a96-b782-fb17eb11b912" (UID: "3cf26616-4656-4a96-b782-fb17eb11b912"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.956631 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3cf26616-4656-4a96-b782-fb17eb11b912" (UID: "3cf26616-4656-4a96-b782-fb17eb11b912"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:16:35 crc kubenswrapper[4926]: I1007 21:16:35.956775 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3cf26616-4656-4a96-b782-fb17eb11b912" (UID: "3cf26616-4656-4a96-b782-fb17eb11b912"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.054248 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.054273 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.054284 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cf26616-4656-4a96-b782-fb17eb11b912-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.759250 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8685f8ddff-w54pw" Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.761761 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l8s8j" event={"ID":"4c6aa769-924b-42ea-b9f2-76f2d77abe3a","Type":"ContainerStarted","Data":"332c5791f8740dd0c6f97e584f92d04ad29eb986458d02666f1d29ffbd04864e"} Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.761816 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l8s8j" event={"ID":"4c6aa769-924b-42ea-b9f2-76f2d77abe3a","Type":"ContainerStarted","Data":"f4d2639b9648258d27c49e23b7083716ebedd326d3516928032dc0083dc97669"} Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.785685 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-l8s8j" podStartSLOduration=2.785671161 podStartE2EDuration="2.785671161s" podCreationTimestamp="2025-10-07 21:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:36.783985672 +0000 UTC m=+1246.822566822" watchObservedRunningTime="2025-10-07 21:16:36.785671161 +0000 UTC m=+1246.824252311" Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.821353 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8685f8ddff-w54pw"] Oct 07 21:16:36 crc kubenswrapper[4926]: I1007 21:16:36.834865 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8685f8ddff-w54pw"] Oct 07 21:16:38 crc kubenswrapper[4926]: I1007 21:16:38.693299 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" path="/var/lib/kubelet/pods/3cf26616-4656-4a96-b782-fb17eb11b912/volumes" Oct 07 21:16:41 crc kubenswrapper[4926]: I1007 21:16:41.020691 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:16:41 crc kubenswrapper[4926]: I1007 21:16:41.021692 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:16:41 crc kubenswrapper[4926]: I1007 21:16:41.837948 4926 generic.go:334] "Generic (PLEG): container finished" podID="4c6aa769-924b-42ea-b9f2-76f2d77abe3a" containerID="332c5791f8740dd0c6f97e584f92d04ad29eb986458d02666f1d29ffbd04864e" exitCode=0 Oct 07 21:16:41 crc kubenswrapper[4926]: I1007 21:16:41.838010 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l8s8j" event={"ID":"4c6aa769-924b-42ea-b9f2-76f2d77abe3a","Type":"ContainerDied","Data":"332c5791f8740dd0c6f97e584f92d04ad29eb986458d02666f1d29ffbd04864e"} Oct 07 21:16:42 crc kubenswrapper[4926]: I1007 21:16:42.046436 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.222:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:42 crc kubenswrapper[4926]: I1007 21:16:42.048078 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.222:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.334833 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.426938 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-combined-ca-bundle\") pod \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.427043 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ksbx\" (UniqueName: \"kubernetes.io/projected/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-kube-api-access-9ksbx\") pod \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.427157 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-scripts\") pod \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.427220 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-config-data\") pod \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\" (UID: \"4c6aa769-924b-42ea-b9f2-76f2d77abe3a\") " Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.433387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-kube-api-access-9ksbx" (OuterVolumeSpecName: "kube-api-access-9ksbx") pod "4c6aa769-924b-42ea-b9f2-76f2d77abe3a" (UID: "4c6aa769-924b-42ea-b9f2-76f2d77abe3a"). InnerVolumeSpecName "kube-api-access-9ksbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.439639 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-scripts" (OuterVolumeSpecName: "scripts") pod "4c6aa769-924b-42ea-b9f2-76f2d77abe3a" (UID: "4c6aa769-924b-42ea-b9f2-76f2d77abe3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.471535 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c6aa769-924b-42ea-b9f2-76f2d77abe3a" (UID: "4c6aa769-924b-42ea-b9f2-76f2d77abe3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.472216 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-config-data" (OuterVolumeSpecName: "config-data") pod "4c6aa769-924b-42ea-b9f2-76f2d77abe3a" (UID: "4c6aa769-924b-42ea-b9f2-76f2d77abe3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.530815 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.530869 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ksbx\" (UniqueName: \"kubernetes.io/projected/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-kube-api-access-9ksbx\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.530894 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.530914 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6aa769-924b-42ea-b9f2-76f2d77abe3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.874228 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l8s8j" event={"ID":"4c6aa769-924b-42ea-b9f2-76f2d77abe3a","Type":"ContainerDied","Data":"f4d2639b9648258d27c49e23b7083716ebedd326d3516928032dc0083dc97669"} Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.874305 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4d2639b9648258d27c49e23b7083716ebedd326d3516928032dc0083dc97669" Oct 07 21:16:43 crc kubenswrapper[4926]: I1007 21:16:43.874410 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l8s8j" Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.065161 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.065648 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-log" containerID="cri-o://c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f" gracePeriod=30 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.065754 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-api" containerID="cri-o://a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd" gracePeriod=30 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.079289 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.079583 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" containerName="nova-scheduler-scheduler" containerID="cri-o://7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" gracePeriod=30 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.117939 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.118696 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-log" containerID="cri-o://87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a" gracePeriod=30 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.118773 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-metadata" containerID="cri-o://d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516" gracePeriod=30 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.890669 4926 generic.go:334] "Generic (PLEG): container finished" podID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerID="87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a" exitCode=143 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.890756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35df0b0f-c822-4229-9c0b-f132d8ae09ee","Type":"ContainerDied","Data":"87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a"} Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.892479 4926 generic.go:334] "Generic (PLEG): container finished" podID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerID="c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f" exitCode=143 Oct 07 21:16:44 crc kubenswrapper[4926]: I1007 21:16:44.892514 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"000f063a-c62d-48dc-92a0-b5b6f317d60b","Type":"ContainerDied","Data":"c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f"} Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.511551 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.517823 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.578710 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-nova-metadata-tls-certs\") pod \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.579031 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-combined-ca-bundle\") pod \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.579161 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-internal-tls-certs\") pod \"000f063a-c62d-48dc-92a0-b5b6f317d60b\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.579309 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-config-data\") pod \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.579620 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000f063a-c62d-48dc-92a0-b5b6f317d60b-logs\") pod \"000f063a-c62d-48dc-92a0-b5b6f317d60b\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580044 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-public-tls-certs\") pod \"000f063a-c62d-48dc-92a0-b5b6f317d60b\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580116 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000f063a-c62d-48dc-92a0-b5b6f317d60b-logs" (OuterVolumeSpecName: "logs") pod "000f063a-c62d-48dc-92a0-b5b6f317d60b" (UID: "000f063a-c62d-48dc-92a0-b5b6f317d60b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580338 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-config-data\") pod \"000f063a-c62d-48dc-92a0-b5b6f317d60b\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580476 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35df0b0f-c822-4229-9c0b-f132d8ae09ee-logs\") pod \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580579 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-combined-ca-bundle\") pod \"000f063a-c62d-48dc-92a0-b5b6f317d60b\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580733 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsm4l\" (UniqueName: \"kubernetes.io/projected/000f063a-c62d-48dc-92a0-b5b6f317d60b-kube-api-access-jsm4l\") pod \"000f063a-c62d-48dc-92a0-b5b6f317d60b\" (UID: \"000f063a-c62d-48dc-92a0-b5b6f317d60b\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580845 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6wwp\" (UniqueName: \"kubernetes.io/projected/35df0b0f-c822-4229-9c0b-f132d8ae09ee-kube-api-access-n6wwp\") pod \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\" (UID: \"35df0b0f-c822-4229-9c0b-f132d8ae09ee\") " Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.580931 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35df0b0f-c822-4229-9c0b-f132d8ae09ee-logs" (OuterVolumeSpecName: "logs") pod "35df0b0f-c822-4229-9c0b-f132d8ae09ee" (UID: "35df0b0f-c822-4229-9c0b-f132d8ae09ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.581575 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000f063a-c62d-48dc-92a0-b5b6f317d60b-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.581710 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35df0b0f-c822-4229-9c0b-f132d8ae09ee-logs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.609530 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35df0b0f-c822-4229-9c0b-f132d8ae09ee-kube-api-access-n6wwp" (OuterVolumeSpecName: "kube-api-access-n6wwp") pod "35df0b0f-c822-4229-9c0b-f132d8ae09ee" (UID: "35df0b0f-c822-4229-9c0b-f132d8ae09ee"). InnerVolumeSpecName "kube-api-access-n6wwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.609612 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000f063a-c62d-48dc-92a0-b5b6f317d60b-kube-api-access-jsm4l" (OuterVolumeSpecName: "kube-api-access-jsm4l") pod "000f063a-c62d-48dc-92a0-b5b6f317d60b" (UID: "000f063a-c62d-48dc-92a0-b5b6f317d60b"). InnerVolumeSpecName "kube-api-access-jsm4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.645417 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35df0b0f-c822-4229-9c0b-f132d8ae09ee" (UID: "35df0b0f-c822-4229-9c0b-f132d8ae09ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.656404 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "000f063a-c62d-48dc-92a0-b5b6f317d60b" (UID: "000f063a-c62d-48dc-92a0-b5b6f317d60b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: E1007 21:16:45.671362 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.676425 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-config-data" (OuterVolumeSpecName: "config-data") pod "000f063a-c62d-48dc-92a0-b5b6f317d60b" (UID: "000f063a-c62d-48dc-92a0-b5b6f317d60b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.697654 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.697683 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsm4l\" (UniqueName: \"kubernetes.io/projected/000f063a-c62d-48dc-92a0-b5b6f317d60b-kube-api-access-jsm4l\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.697692 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6wwp\" (UniqueName: \"kubernetes.io/projected/35df0b0f-c822-4229-9c0b-f132d8ae09ee-kube-api-access-n6wwp\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.697701 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.697709 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: E1007 21:16:45.700360 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 21:16:45 crc kubenswrapper[4926]: E1007 21:16:45.734632 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 21:16:45 crc kubenswrapper[4926]: E1007 21:16:45.734707 4926 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" containerName="nova-scheduler-scheduler" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.802597 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-config-data" (OuterVolumeSpecName: "config-data") pod "35df0b0f-c822-4229-9c0b-f132d8ae09ee" (UID: "35df0b0f-c822-4229-9c0b-f132d8ae09ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.803758 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.816703 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "35df0b0f-c822-4229-9c0b-f132d8ae09ee" (UID: "35df0b0f-c822-4229-9c0b-f132d8ae09ee"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.826338 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "000f063a-c62d-48dc-92a0-b5b6f317d60b" (UID: "000f063a-c62d-48dc-92a0-b5b6f317d60b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.857364 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "000f063a-c62d-48dc-92a0-b5b6f317d60b" (UID: "000f063a-c62d-48dc-92a0-b5b6f317d60b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.906181 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/35df0b0f-c822-4229-9c0b-f132d8ae09ee-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.906293 4926 generic.go:334] "Generic (PLEG): container finished" podID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerID="a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd" exitCode=0 Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.906340 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"000f063a-c62d-48dc-92a0-b5b6f317d60b","Type":"ContainerDied","Data":"a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd"} Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.906365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"000f063a-c62d-48dc-92a0-b5b6f317d60b","Type":"ContainerDied","Data":"d4476f5b51e499504309d1eb4ba56059882eb0d3bca36bc23ee62de924486eb7"} Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.906381 4926 scope.go:117] "RemoveContainer" containerID="a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.906479 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.914272 4926 generic.go:334] "Generic (PLEG): container finished" podID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerID="d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516" exitCode=0 Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.914328 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35df0b0f-c822-4229-9c0b-f132d8ae09ee","Type":"ContainerDied","Data":"d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516"} Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.914351 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.914363 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35df0b0f-c822-4229-9c0b-f132d8ae09ee","Type":"ContainerDied","Data":"5990f05ce881eb0cc2d0ac5b3af5f847b9a8c6d8073e407569bcd09ef815823a"} Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.918387 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.918409 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000f063a-c62d-48dc-92a0-b5b6f317d60b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.945224 4926 scope.go:117] "RemoveContainer" containerID="c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.945472 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.961554 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.971519 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.973449 4926 scope.go:117] "RemoveContainer" containerID="a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd" Oct 07 21:16:45 crc kubenswrapper[4926]: E1007 21:16:45.977814 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd\": container with ID starting with a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd not found: ID does not exist" containerID="a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.977933 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd"} err="failed to get container status \"a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd\": rpc error: code = NotFound desc = could not find container \"a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd\": container with ID starting with a45f9bb2a005ad4f467d52070b5dbdafa3a51a8c38e9a51c1aeccf9fae30a2cd not found: ID does not exist" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.978008 4926 scope.go:117] "RemoveContainer" containerID="c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f" Oct 07 21:16:45 crc kubenswrapper[4926]: E1007 21:16:45.978908 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f\": container with ID starting with c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f not found: ID does not exist" containerID="c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.978967 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f"} err="failed to get container status \"c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f\": rpc error: code = NotFound desc = could not find container \"c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f\": container with ID starting with c1b36b5ec1355ed64ff63116b1368bf4db5d850dd6f15e74116c2992218d0f0f not found: ID does not exist" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.979001 4926 scope.go:117] "RemoveContainer" containerID="d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516" Oct 07 21:16:45 crc kubenswrapper[4926]: I1007 21:16:45.984985 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.003578 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004101 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" containerName="init" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004124 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" containerName="init" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004145 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-metadata" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004153 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-metadata" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004164 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-api" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004170 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-api" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004185 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" containerName="dnsmasq-dns" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004214 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" containerName="dnsmasq-dns" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004234 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-log" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004243 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-log" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004271 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-log" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004279 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-log" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.004294 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6aa769-924b-42ea-b9f2-76f2d77abe3a" containerName="nova-manage" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004301 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6aa769-924b-42ea-b9f2-76f2d77abe3a" containerName="nova-manage" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004505 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-metadata" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004529 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-log" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004548 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf26616-4656-4a96-b782-fb17eb11b912" containerName="dnsmasq-dns" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004562 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" containerName="nova-api-api" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004573 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6aa769-924b-42ea-b9f2-76f2d77abe3a" containerName="nova-manage" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.004585 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" containerName="nova-metadata-log" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.005882 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.010088 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.010635 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.010851 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.011895 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.015115 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.015468 4926 scope.go:117] "RemoveContainer" containerID="87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.016518 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.017585 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.022536 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.033580 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.045788 4926 scope.go:117] "RemoveContainer" containerID="d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.046354 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516\": container with ID starting with d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516 not found: ID does not exist" containerID="d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.046445 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516"} err="failed to get container status \"d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516\": rpc error: code = NotFound desc = could not find container \"d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516\": container with ID starting with d804c51d313d9d6023b5b602c6e9df8c0eba705f0f2701632105952f5f0cb516 not found: ID does not exist" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.046520 4926 scope.go:117] "RemoveContainer" containerID="87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a" Oct 07 21:16:46 crc kubenswrapper[4926]: E1007 21:16:46.047052 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a\": container with ID starting with 87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a not found: ID does not exist" containerID="87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.047107 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a"} err="failed to get container status \"87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a\": rpc error: code = NotFound desc = could not find container \"87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a\": container with ID starting with 87eb86a760a939d1f4572178494b4a2186e501bd2cb569752cbbb429bd2bed1a not found: ID does not exist" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.122653 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.122724 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-public-tls-certs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.122751 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64mrn\" (UniqueName: \"kubernetes.io/projected/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-kube-api-access-64mrn\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.122902 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-config-data\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.123175 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.123320 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-config-data\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.123759 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnckg\" (UniqueName: \"kubernetes.io/projected/08044ebc-ce53-4476-9884-83792098080d-kube-api-access-pnckg\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.123808 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.123886 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-logs\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.123986 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.124111 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08044ebc-ce53-4476-9884-83792098080d-logs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226024 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnckg\" (UniqueName: \"kubernetes.io/projected/08044ebc-ce53-4476-9884-83792098080d-kube-api-access-pnckg\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226136 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-logs\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226171 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226244 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08044ebc-ce53-4476-9884-83792098080d-logs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226292 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226323 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-public-tls-certs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226351 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64mrn\" (UniqueName: \"kubernetes.io/projected/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-kube-api-access-64mrn\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226408 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-config-data\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226451 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.226485 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-config-data\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.227538 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08044ebc-ce53-4476-9884-83792098080d-logs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.227868 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-logs\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.230444 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.231732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.232401 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-config-data\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.232436 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.233327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-public-tls-certs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.234544 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08044ebc-ce53-4476-9884-83792098080d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.237010 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-config-data\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.244007 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64mrn\" (UniqueName: \"kubernetes.io/projected/4496af78-6ebe-4aec-a04e-dcf4e8f446eb-kube-api-access-64mrn\") pod \"nova-metadata-0\" (UID: \"4496af78-6ebe-4aec-a04e-dcf4e8f446eb\") " pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.250965 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnckg\" (UniqueName: \"kubernetes.io/projected/08044ebc-ce53-4476-9884-83792098080d-kube-api-access-pnckg\") pod \"nova-api-0\" (UID: \"08044ebc-ce53-4476-9884-83792098080d\") " pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.327650 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.346493 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.691339 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000f063a-c62d-48dc-92a0-b5b6f317d60b" path="/var/lib/kubelet/pods/000f063a-c62d-48dc-92a0-b5b6f317d60b/volumes" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.692300 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35df0b0f-c822-4229-9c0b-f132d8ae09ee" path="/var/lib/kubelet/pods/35df0b0f-c822-4229-9c0b-f132d8ae09ee/volumes" Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.864869 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 21:16:46 crc kubenswrapper[4926]: W1007 21:16:46.870622 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08044ebc_ce53_4476_9884_83792098080d.slice/crio-3d8b6d052ddb6d81a40723cac43ba2eae195ee2db185ba97296271ed78fd6fc2 WatchSource:0}: Error finding container 3d8b6d052ddb6d81a40723cac43ba2eae195ee2db185ba97296271ed78fd6fc2: Status 404 returned error can't find the container with id 3d8b6d052ddb6d81a40723cac43ba2eae195ee2db185ba97296271ed78fd6fc2 Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.929542 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08044ebc-ce53-4476-9884-83792098080d","Type":"ContainerStarted","Data":"3d8b6d052ddb6d81a40723cac43ba2eae195ee2db185ba97296271ed78fd6fc2"} Oct 07 21:16:46 crc kubenswrapper[4926]: I1007 21:16:46.958648 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 21:16:47 crc kubenswrapper[4926]: I1007 21:16:47.958512 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08044ebc-ce53-4476-9884-83792098080d","Type":"ContainerStarted","Data":"ff6282a7169950ce086c95f2d1cb628e4e4d53b3dd5f3fa7aad1b920d2835b9f"} Oct 07 21:16:47 crc kubenswrapper[4926]: I1007 21:16:47.959107 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08044ebc-ce53-4476-9884-83792098080d","Type":"ContainerStarted","Data":"bb86156b3af609d8993f75d9ae6a930cd277f7609d588ff7db7692938d290aa3"} Oct 07 21:16:47 crc kubenswrapper[4926]: I1007 21:16:47.965076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4496af78-6ebe-4aec-a04e-dcf4e8f446eb","Type":"ContainerStarted","Data":"4eee5dd9ae46fb2d5c20d6d64318b05d4a89e9515c454b0e54fb29bcef429cab"} Oct 07 21:16:47 crc kubenswrapper[4926]: I1007 21:16:47.965616 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4496af78-6ebe-4aec-a04e-dcf4e8f446eb","Type":"ContainerStarted","Data":"7a1bd46c6b40650da102ce85deb094b3112e58b7885ae669ff806dce2042ae69"} Oct 07 21:16:47 crc kubenswrapper[4926]: I1007 21:16:47.965628 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4496af78-6ebe-4aec-a04e-dcf4e8f446eb","Type":"ContainerStarted","Data":"b50fb413ea326b129441e9610393c80f1a22de57dcc119a3bf3b1c16035a8f0a"} Oct 07 21:16:47 crc kubenswrapper[4926]: I1007 21:16:47.980935 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.980916066 podStartE2EDuration="2.980916066s" podCreationTimestamp="2025-10-07 21:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:47.978061203 +0000 UTC m=+1258.016642373" watchObservedRunningTime="2025-10-07 21:16:47.980916066 +0000 UTC m=+1258.019497216" Oct 07 21:16:48 crc kubenswrapper[4926]: I1007 21:16:48.005139 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.005090089 podStartE2EDuration="3.005090089s" podCreationTimestamp="2025-10-07 21:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:48.003541814 +0000 UTC m=+1258.042122974" watchObservedRunningTime="2025-10-07 21:16:48.005090089 +0000 UTC m=+1258.043671249" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.695148 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.805504 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-config-data\") pod \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.805738 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-combined-ca-bundle\") pod \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.805861 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nbwd\" (UniqueName: \"kubernetes.io/projected/cd696bf0-2ccd-45b4-a498-698ad9406dc4-kube-api-access-4nbwd\") pod \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\" (UID: \"cd696bf0-2ccd-45b4-a498-698ad9406dc4\") " Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.810574 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd696bf0-2ccd-45b4-a498-698ad9406dc4-kube-api-access-4nbwd" (OuterVolumeSpecName: "kube-api-access-4nbwd") pod "cd696bf0-2ccd-45b4-a498-698ad9406dc4" (UID: "cd696bf0-2ccd-45b4-a498-698ad9406dc4"). InnerVolumeSpecName "kube-api-access-4nbwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.839550 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd696bf0-2ccd-45b4-a498-698ad9406dc4" (UID: "cd696bf0-2ccd-45b4-a498-698ad9406dc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.839578 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-config-data" (OuterVolumeSpecName: "config-data") pod "cd696bf0-2ccd-45b4-a498-698ad9406dc4" (UID: "cd696bf0-2ccd-45b4-a498-698ad9406dc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.908562 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.908597 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd696bf0-2ccd-45b4-a498-698ad9406dc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.908611 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nbwd\" (UniqueName: \"kubernetes.io/projected/cd696bf0-2ccd-45b4-a498-698ad9406dc4-kube-api-access-4nbwd\") on node \"crc\" DevicePath \"\"" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.994681 4926 generic.go:334] "Generic (PLEG): container finished" podID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" exitCode=0 Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.994742 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd696bf0-2ccd-45b4-a498-698ad9406dc4","Type":"ContainerDied","Data":"7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26"} Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.994782 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.994818 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd696bf0-2ccd-45b4-a498-698ad9406dc4","Type":"ContainerDied","Data":"80afd5b03259761ba992d4fc4d5f01ce2f9666020e5a3afbcf3ebf390429e04c"} Oct 07 21:16:49 crc kubenswrapper[4926]: I1007 21:16:49.994848 4926 scope.go:117] "RemoveContainer" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.021300 4926 scope.go:117] "RemoveContainer" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" Oct 07 21:16:50 crc kubenswrapper[4926]: E1007 21:16:50.022375 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26\": container with ID starting with 7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26 not found: ID does not exist" containerID="7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.022427 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26"} err="failed to get container status \"7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26\": rpc error: code = NotFound desc = could not find container \"7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26\": container with ID starting with 7dd063c503321e6657160ca9c200f94587292f07f67451f991bedafa97da3a26 not found: ID does not exist" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.042468 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.051293 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.058246 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:50 crc kubenswrapper[4926]: E1007 21:16:50.058719 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" containerName="nova-scheduler-scheduler" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.058737 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" containerName="nova-scheduler-scheduler" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.058971 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" containerName="nova-scheduler-scheduler" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.059728 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.065340 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.068800 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.112416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c2rv\" (UniqueName: \"kubernetes.io/projected/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-kube-api-access-7c2rv\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.112466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.112636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-config-data\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.214229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c2rv\" (UniqueName: \"kubernetes.io/projected/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-kube-api-access-7c2rv\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.214494 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.214584 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-config-data\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.220234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.222224 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-config-data\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.230063 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c2rv\" (UniqueName: \"kubernetes.io/projected/c5fe2fcd-3215-4278-90fc-3adc5cec14fe-kube-api-access-7c2rv\") pod \"nova-scheduler-0\" (UID: \"c5fe2fcd-3215-4278-90fc-3adc5cec14fe\") " pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.383579 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.693798 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd696bf0-2ccd-45b4-a498-698ad9406dc4" path="/var/lib/kubelet/pods/cd696bf0-2ccd-45b4-a498-698ad9406dc4/volumes" Oct 07 21:16:50 crc kubenswrapper[4926]: I1007 21:16:50.849534 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 21:16:51 crc kubenswrapper[4926]: I1007 21:16:51.011178 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5fe2fcd-3215-4278-90fc-3adc5cec14fe","Type":"ContainerStarted","Data":"37bec8c25711cba8cb54094a8dfc773791e1f8629a1ab11c17068cb49dd87f88"} Oct 07 21:16:51 crc kubenswrapper[4926]: I1007 21:16:51.346822 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 21:16:51 crc kubenswrapper[4926]: I1007 21:16:51.346900 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 21:16:52 crc kubenswrapper[4926]: I1007 21:16:52.023820 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5fe2fcd-3215-4278-90fc-3adc5cec14fe","Type":"ContainerStarted","Data":"cfb9aeac89a1af7955676f73adef1770eb464845cf9c42e3d17859b5895bb46f"} Oct 07 21:16:52 crc kubenswrapper[4926]: I1007 21:16:52.045275 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.045251265 podStartE2EDuration="2.045251265s" podCreationTimestamp="2025-10-07 21:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:16:52.036757088 +0000 UTC m=+1262.075338308" watchObservedRunningTime="2025-10-07 21:16:52.045251265 +0000 UTC m=+1262.083832455" Oct 07 21:16:55 crc kubenswrapper[4926]: I1007 21:16:55.384526 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 21:16:56 crc kubenswrapper[4926]: I1007 21:16:56.327879 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:16:56 crc kubenswrapper[4926]: I1007 21:16:56.328290 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 21:16:56 crc kubenswrapper[4926]: I1007 21:16:56.347835 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 21:16:56 crc kubenswrapper[4926]: I1007 21:16:56.348335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 21:16:57 crc kubenswrapper[4926]: I1007 21:16:57.340418 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="08044ebc-ce53-4476-9884-83792098080d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:57 crc kubenswrapper[4926]: I1007 21:16:57.340425 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="08044ebc-ce53-4476-9884-83792098080d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:57 crc kubenswrapper[4926]: I1007 21:16:57.362465 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4496af78-6ebe-4aec-a04e-dcf4e8f446eb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.226:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 21:16:57 crc kubenswrapper[4926]: I1007 21:16:57.362466 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4496af78-6ebe-4aec-a04e-dcf4e8f446eb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.226:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 21:17:00 crc kubenswrapper[4926]: I1007 21:17:00.383960 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 21:17:00 crc kubenswrapper[4926]: I1007 21:17:00.413571 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 21:17:01 crc kubenswrapper[4926]: I1007 21:17:01.182430 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 21:17:02 crc kubenswrapper[4926]: I1007 21:17:02.071479 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 21:17:03 crc kubenswrapper[4926]: I1007 21:17:03.209785 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:17:03 crc kubenswrapper[4926]: I1007 21:17:03.209848 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.340413 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.341864 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.343567 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.354476 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.356789 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.357326 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 21:17:06 crc kubenswrapper[4926]: I1007 21:17:06.362505 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 21:17:07 crc kubenswrapper[4926]: I1007 21:17:07.207320 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 21:17:07 crc kubenswrapper[4926]: I1007 21:17:07.214939 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 21:17:07 crc kubenswrapper[4926]: I1007 21:17:07.227183 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 21:17:15 crc kubenswrapper[4926]: I1007 21:17:15.204746 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:17:15 crc kubenswrapper[4926]: I1007 21:17:15.949172 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:17:18 crc kubenswrapper[4926]: I1007 21:17:18.515322 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="rabbitmq" containerID="cri-o://e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f" gracePeriod=604797 Oct 07 21:17:19 crc kubenswrapper[4926]: I1007 21:17:19.202408 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="rabbitmq" containerID="cri-o://d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba" gracePeriod=604797 Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.256775 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.349652 4926 generic.go:334] "Generic (PLEG): container finished" podID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerID="e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f" exitCode=0 Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.349697 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0","Type":"ContainerDied","Data":"e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f"} Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.349723 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0","Type":"ContainerDied","Data":"c4f75ba6a03bce10213bd90cfd506bb75e16eae00d638627ad4eb4e2882a3eb1"} Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.349739 4926 scope.go:117] "RemoveContainer" containerID="e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.349870 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375444 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375501 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-config-data\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375549 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-confd\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375655 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-tls\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375680 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5lls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-kube-api-access-z5lls\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375695 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-plugins-conf\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375751 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-erlang-cookie-secret\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375848 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-pod-info\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375866 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-server-conf\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375887 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-plugins\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.375917 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-erlang-cookie\") pod \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\" (UID: \"a9d51a12-f1e2-4b2e-a5d7-51474b31edb0\") " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.376767 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.377132 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.379221 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.385959 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-pod-info" (OuterVolumeSpecName: "pod-info") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.387791 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.389820 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.395094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.407508 4926 scope.go:117] "RemoveContainer" containerID="e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.407615 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-kube-api-access-z5lls" (OuterVolumeSpecName: "kube-api-access-z5lls") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "kube-api-access-z5lls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.442100 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-config-data" (OuterVolumeSpecName: "config-data") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480145 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480187 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5lls\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-kube-api-access-z5lls\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480217 4926 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480229 4926 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480241 4926 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480252 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480264 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480289 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.480301 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.506822 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.510055 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-server-conf" (OuterVolumeSpecName: "server-conf") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.546025 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" (UID: "a9d51a12-f1e2-4b2e-a5d7-51474b31edb0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.576332 4926 scope.go:117] "RemoveContainer" containerID="e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f" Oct 07 21:17:20 crc kubenswrapper[4926]: E1007 21:17:20.576861 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f\": container with ID starting with e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f not found: ID does not exist" containerID="e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.576898 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f"} err="failed to get container status \"e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f\": rpc error: code = NotFound desc = could not find container \"e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f\": container with ID starting with e8aa3a1425ae9e396c37a490463f273f82fa1ebd99a3f4548dfc2661f352cf4f not found: ID does not exist" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.576924 4926 scope.go:117] "RemoveContainer" containerID="e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad" Oct 07 21:17:20 crc kubenswrapper[4926]: E1007 21:17:20.577348 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad\": container with ID starting with e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad not found: ID does not exist" containerID="e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.577388 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad"} err="failed to get container status \"e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad\": rpc error: code = NotFound desc = could not find container \"e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad\": container with ID starting with e724f7d8f739f77f21e1df8e0b762a59a381619f5187a0b80184324743e40fad not found: ID does not exist" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.582750 4926 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.582780 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.582789 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.692498 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.712409 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.723264 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:17:20 crc kubenswrapper[4926]: E1007 21:17:20.723665 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="setup-container" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.723684 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="setup-container" Oct 07 21:17:20 crc kubenswrapper[4926]: E1007 21:17:20.723721 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="rabbitmq" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.723730 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="rabbitmq" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.723946 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" containerName="rabbitmq" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.724991 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.729430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.729652 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.729766 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.729870 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lmqmr" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.729975 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.730079 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.730225 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.746801 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.785769 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhbtq\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-kube-api-access-dhbtq\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.785813 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.785843 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.785864 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.785906 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.785934 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-config-data\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.786021 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.786044 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.786076 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.786091 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.786112 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.888665 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhbtq\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-kube-api-access-dhbtq\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.888788 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.888852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.888881 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.888961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.889024 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-config-data\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.889126 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.889174 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.889241 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.889266 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.889319 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.890654 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.890763 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-config-data\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.890932 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.892236 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.894325 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.894354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.895765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.896240 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.897968 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.898670 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.911742 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhbtq\" (UniqueName: \"kubernetes.io/projected/9052310e-1ddc-4b9b-a3bc-ced3dc97ab21-kube-api-access-dhbtq\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:20 crc kubenswrapper[4926]: I1007 21:17:20.931724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21\") " pod="openstack/rabbitmq-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.005092 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.046453 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091770 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-tls\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091832 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6hvx\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-kube-api-access-f6hvx\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-pod-info\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091878 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-plugins\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-server-conf\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091946 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-confd\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.091963 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-erlang-cookie-secret\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.092065 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-erlang-cookie\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.092167 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-config-data\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.092277 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-plugins-conf\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.092361 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\" (UID: \"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57\") " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.092883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.094524 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.094873 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.100545 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.100820 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-pod-info" (OuterVolumeSpecName: "pod-info") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.128390 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.128522 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-kube-api-access-f6hvx" (OuterVolumeSpecName: "kube-api-access-f6hvx") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "kube-api-access-f6hvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.143545 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195672 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6hvx\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-kube-api-access-f6hvx\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195716 4926 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195729 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195743 4926 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195755 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195765 4926 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195799 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.195812 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.243798 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-config-data" (OuterVolumeSpecName: "config-data") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.259966 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.310323 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.310570 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.328122 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.361226 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-server-conf" (OuterVolumeSpecName: "server-conf") pod "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" (UID: "59e27b2d-9c0d-46a6-8be6-eee9cdf04c57"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.367907 4926 generic.go:334] "Generic (PLEG): container finished" podID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerID="d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba" exitCode=0 Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.368004 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.368045 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57","Type":"ContainerDied","Data":"d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba"} Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.368094 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"59e27b2d-9c0d-46a6-8be6-eee9cdf04c57","Type":"ContainerDied","Data":"92ca6955f5c8830dfdc19799eb080626c55ec8e7d8f9ffceba836b2003e7b1ad"} Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.368114 4926 scope.go:117] "RemoveContainer" containerID="d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.414075 4926 scope.go:117] "RemoveContainer" containerID="aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.414590 4926 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.414619 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.432271 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.454188 4926 scope.go:117] "RemoveContainer" containerID="d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba" Oct 07 21:17:21 crc kubenswrapper[4926]: E1007 21:17:21.455508 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba\": container with ID starting with d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba not found: ID does not exist" containerID="d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.455543 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba"} err="failed to get container status \"d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba\": rpc error: code = NotFound desc = could not find container \"d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba\": container with ID starting with d46cd0dd99bac7c6d443fafd4e6b8435f2d6026164e104b5d092ed721039b7ba not found: ID does not exist" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.455572 4926 scope.go:117] "RemoveContainer" containerID="aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231" Oct 07 21:17:21 crc kubenswrapper[4926]: E1007 21:17:21.457116 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231\": container with ID starting with aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231 not found: ID does not exist" containerID="aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.457146 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231"} err="failed to get container status \"aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231\": rpc error: code = NotFound desc = could not find container \"aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231\": container with ID starting with aaff461c7549f13fdfdf30e7c85d60f65f915d60fb961a1cf7180f936f977231 not found: ID does not exist" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.458802 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.510226 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:17:21 crc kubenswrapper[4926]: E1007 21:17:21.510684 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="setup-container" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.510704 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="setup-container" Oct 07 21:17:21 crc kubenswrapper[4926]: E1007 21:17:21.510722 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="rabbitmq" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.510728 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="rabbitmq" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.510967 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" containerName="rabbitmq" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.512414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.518627 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.518696 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.518848 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.518943 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5k582" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.519082 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.519227 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.519889 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.537067 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622352 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622389 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622409 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622450 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622475 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8b7q\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-kube-api-access-m8b7q\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622493 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622516 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622550 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622606 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.622633 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.687623 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 21:17:21 crc kubenswrapper[4926]: W1007 21:17:21.692231 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9052310e_1ddc_4b9b_a3bc_ced3dc97ab21.slice/crio-631d2c1a53cd1e9793f3dcbece1385f3f481364109b2ded15949d9928a7db32b WatchSource:0}: Error finding container 631d2c1a53cd1e9793f3dcbece1385f3f481364109b2ded15949d9928a7db32b: Status 404 returned error can't find the container with id 631d2c1a53cd1e9793f3dcbece1385f3f481364109b2ded15949d9928a7db32b Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.724821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725228 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725360 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725523 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725659 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8b7q\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-kube-api-access-m8b7q\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725847 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725876 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.725999 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.726034 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.726183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.726658 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.726668 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.726767 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.726789 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.728853 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.730022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.730131 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.730283 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.741709 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.745870 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8b7q\" (UniqueName: \"kubernetes.io/projected/d5d417c9-ac1d-4b58-9506-bc6dd17a71b1-kube-api-access-m8b7q\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.768669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:21 crc kubenswrapper[4926]: I1007 21:17:21.840119 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:22 crc kubenswrapper[4926]: I1007 21:17:22.288585 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 21:17:22 crc kubenswrapper[4926]: W1007 21:17:22.291656 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5d417c9_ac1d_4b58_9506_bc6dd17a71b1.slice/crio-130ab2a7ecddeb4a0b5f9f46d05bfbf796437498b13407a4ecda680754d09d5c WatchSource:0}: Error finding container 130ab2a7ecddeb4a0b5f9f46d05bfbf796437498b13407a4ecda680754d09d5c: Status 404 returned error can't find the container with id 130ab2a7ecddeb4a0b5f9f46d05bfbf796437498b13407a4ecda680754d09d5c Oct 07 21:17:22 crc kubenswrapper[4926]: I1007 21:17:22.396687 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1","Type":"ContainerStarted","Data":"130ab2a7ecddeb4a0b5f9f46d05bfbf796437498b13407a4ecda680754d09d5c"} Oct 07 21:17:22 crc kubenswrapper[4926]: I1007 21:17:22.397626 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21","Type":"ContainerStarted","Data":"631d2c1a53cd1e9793f3dcbece1385f3f481364109b2ded15949d9928a7db32b"} Oct 07 21:17:22 crc kubenswrapper[4926]: I1007 21:17:22.692038 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e27b2d-9c0d-46a6-8be6-eee9cdf04c57" path="/var/lib/kubelet/pods/59e27b2d-9c0d-46a6-8be6-eee9cdf04c57/volumes" Oct 07 21:17:22 crc kubenswrapper[4926]: I1007 21:17:22.693227 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d51a12-f1e2-4b2e-a5d7-51474b31edb0" path="/var/lib/kubelet/pods/a9d51a12-f1e2-4b2e-a5d7-51474b31edb0/volumes" Oct 07 21:17:24 crc kubenswrapper[4926]: I1007 21:17:24.419240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1","Type":"ContainerStarted","Data":"92b938d857b1e307613d07faae7fbd33fc7fe53b8d6a602a63add4986e80d112"} Oct 07 21:17:24 crc kubenswrapper[4926]: I1007 21:17:24.422386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21","Type":"ContainerStarted","Data":"6f6ae2be9b9fe803269e11f8b3855557c4dc71bcd39e32ca19cf83b1d7beb3c6"} Oct 07 21:17:30 crc kubenswrapper[4926]: I1007 21:17:30.982170 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-867bc666f9-58g2f"] Oct 07 21:17:30 crc kubenswrapper[4926]: I1007 21:17:30.985286 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:30 crc kubenswrapper[4926]: I1007 21:17:30.989623 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.007343 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867bc666f9-58g2f"] Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.131241 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-config\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.131325 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-nb\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.131549 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-openstack-edpm-ipam\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.131732 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmc66\" (UniqueName: \"kubernetes.io/projected/0a67502a-9033-4674-a475-3e5d3bb3d379-kube-api-access-gmc66\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.131886 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-svc\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.131992 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-sb\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.132162 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-swift-storage-0\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.234930 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-nb\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.235052 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-openstack-edpm-ipam\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.235107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmc66\" (UniqueName: \"kubernetes.io/projected/0a67502a-9033-4674-a475-3e5d3bb3d379-kube-api-access-gmc66\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.235142 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-svc\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.235187 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-sb\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.235244 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-swift-storage-0\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.235282 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-config\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.236421 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-config\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.236504 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-nb\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.237145 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-svc\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.238724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-sb\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.239008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-openstack-edpm-ipam\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.239166 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-swift-storage-0\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.258845 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmc66\" (UniqueName: \"kubernetes.io/projected/0a67502a-9033-4674-a475-3e5d3bb3d379-kube-api-access-gmc66\") pod \"dnsmasq-dns-867bc666f9-58g2f\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.308647 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:31 crc kubenswrapper[4926]: I1007 21:17:31.753300 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867bc666f9-58g2f"] Oct 07 21:17:32 crc kubenswrapper[4926]: I1007 21:17:32.522683 4926 generic.go:334] "Generic (PLEG): container finished" podID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerID="126021cfb71624ff4bc09529a1b34ee8edd32c30ae300727b73d3f2d1bae712a" exitCode=0 Oct 07 21:17:32 crc kubenswrapper[4926]: I1007 21:17:32.522756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" event={"ID":"0a67502a-9033-4674-a475-3e5d3bb3d379","Type":"ContainerDied","Data":"126021cfb71624ff4bc09529a1b34ee8edd32c30ae300727b73d3f2d1bae712a"} Oct 07 21:17:32 crc kubenswrapper[4926]: I1007 21:17:32.522941 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" event={"ID":"0a67502a-9033-4674-a475-3e5d3bb3d379","Type":"ContainerStarted","Data":"034abbaa010c04aa94c4f428f67dd01ca9c3d4dceb69972fc07c29565a853f44"} Oct 07 21:17:33 crc kubenswrapper[4926]: I1007 21:17:33.209158 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:17:33 crc kubenswrapper[4926]: I1007 21:17:33.209731 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:17:33 crc kubenswrapper[4926]: I1007 21:17:33.537500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" event={"ID":"0a67502a-9033-4674-a475-3e5d3bb3d379","Type":"ContainerStarted","Data":"092780f8b17ab628d3e2f5be0f078f8e4e57bd9fc15168513a45cc8577deb6d5"} Oct 07 21:17:33 crc kubenswrapper[4926]: I1007 21:17:33.537829 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:33 crc kubenswrapper[4926]: I1007 21:17:33.580306 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" podStartSLOduration=3.580271105 podStartE2EDuration="3.580271105s" podCreationTimestamp="2025-10-07 21:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:17:33.572423737 +0000 UTC m=+1303.611004917" watchObservedRunningTime="2025-10-07 21:17:33.580271105 +0000 UTC m=+1303.618852265" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.310648 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.411380 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f9cb5965-h6hqc"] Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.411838 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerName="dnsmasq-dns" containerID="cri-o://87c7b36a97a6de76ce763e116a6f3dab812136a114e90a2387da5730a3c6283e" gracePeriod=10 Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.667872 4926 generic.go:334] "Generic (PLEG): container finished" podID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerID="87c7b36a97a6de76ce763e116a6f3dab812136a114e90a2387da5730a3c6283e" exitCode=0 Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.667936 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" event={"ID":"3079e3d9-da18-49f6-b29b-275d42b4885c","Type":"ContainerDied","Data":"87c7b36a97a6de76ce763e116a6f3dab812136a114e90a2387da5730a3c6283e"} Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.672360 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c8d6588b7-27rms"] Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.677824 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.698814 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8d6588b7-27rms"] Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.781381 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-dns-svc\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.782356 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.782497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.782532 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6b44\" (UniqueName: \"kubernetes.io/projected/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-kube-api-access-l6b44\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.782600 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-config\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.782700 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.782756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-openstack-edpm-ipam\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.888620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.888758 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.888791 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6b44\" (UniqueName: \"kubernetes.io/projected/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-kube-api-access-l6b44\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.888845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-config\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.888903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.888944 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-openstack-edpm-ipam\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.889033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-dns-svc\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.889777 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.890109 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-dns-svc\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.890666 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-config\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.890987 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.891015 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-openstack-edpm-ipam\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.891147 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:41 crc kubenswrapper[4926]: I1007 21:17:41.914238 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6b44\" (UniqueName: \"kubernetes.io/projected/f8f8c2fb-1675-4abc-8ae2-dbe68370e347-kube-api-access-l6b44\") pod \"dnsmasq-dns-5c8d6588b7-27rms\" (UID: \"f8f8c2fb-1675-4abc-8ae2-dbe68370e347\") " pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.005935 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.550786 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8d6588b7-27rms"] Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.700113 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" event={"ID":"f8f8c2fb-1675-4abc-8ae2-dbe68370e347","Type":"ContainerStarted","Data":"cfda5111aeb974f2562016289859c0eb73d38ff59529f719de0d45bf524cfc39"} Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.700150 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" event={"ID":"3079e3d9-da18-49f6-b29b-275d42b4885c","Type":"ContainerDied","Data":"10d2102e4b71bc3230485a81b62873dd734db2ee7b910542c0f4d265d0393f81"} Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.700164 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10d2102e4b71bc3230485a81b62873dd734db2ee7b910542c0f4d265d0393f81" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.725216 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.816693 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-nb\") pod \"3079e3d9-da18-49f6-b29b-275d42b4885c\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.816740 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-sb\") pod \"3079e3d9-da18-49f6-b29b-275d42b4885c\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.816787 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-swift-storage-0\") pod \"3079e3d9-da18-49f6-b29b-275d42b4885c\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.816817 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-config\") pod \"3079e3d9-da18-49f6-b29b-275d42b4885c\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.816842 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-svc\") pod \"3079e3d9-da18-49f6-b29b-275d42b4885c\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.816899 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpzg5\" (UniqueName: \"kubernetes.io/projected/3079e3d9-da18-49f6-b29b-275d42b4885c-kube-api-access-mpzg5\") pod \"3079e3d9-da18-49f6-b29b-275d42b4885c\" (UID: \"3079e3d9-da18-49f6-b29b-275d42b4885c\") " Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.821182 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3079e3d9-da18-49f6-b29b-275d42b4885c-kube-api-access-mpzg5" (OuterVolumeSpecName: "kube-api-access-mpzg5") pod "3079e3d9-da18-49f6-b29b-275d42b4885c" (UID: "3079e3d9-da18-49f6-b29b-275d42b4885c"). InnerVolumeSpecName "kube-api-access-mpzg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.866683 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3079e3d9-da18-49f6-b29b-275d42b4885c" (UID: "3079e3d9-da18-49f6-b29b-275d42b4885c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.867828 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3079e3d9-da18-49f6-b29b-275d42b4885c" (UID: "3079e3d9-da18-49f6-b29b-275d42b4885c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.874807 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3079e3d9-da18-49f6-b29b-275d42b4885c" (UID: "3079e3d9-da18-49f6-b29b-275d42b4885c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.875442 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-config" (OuterVolumeSpecName: "config") pod "3079e3d9-da18-49f6-b29b-275d42b4885c" (UID: "3079e3d9-da18-49f6-b29b-275d42b4885c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.914383 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3079e3d9-da18-49f6-b29b-275d42b4885c" (UID: "3079e3d9-da18-49f6-b29b-275d42b4885c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.920947 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.921008 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.921022 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.921033 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.921044 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3079e3d9-da18-49f6-b29b-275d42b4885c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:42 crc kubenswrapper[4926]: I1007 21:17:42.921057 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpzg5\" (UniqueName: \"kubernetes.io/projected/3079e3d9-da18-49f6-b29b-275d42b4885c-kube-api-access-mpzg5\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:43 crc kubenswrapper[4926]: I1007 21:17:43.709237 4926 generic.go:334] "Generic (PLEG): container finished" podID="f8f8c2fb-1675-4abc-8ae2-dbe68370e347" containerID="10cb1811294b98f36229b594eb4b631d0de9e984be8bc388d78725fce94a2c5a" exitCode=0 Oct 07 21:17:43 crc kubenswrapper[4926]: I1007 21:17:43.709544 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f9cb5965-h6hqc" Oct 07 21:17:43 crc kubenswrapper[4926]: I1007 21:17:43.711345 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" event={"ID":"f8f8c2fb-1675-4abc-8ae2-dbe68370e347","Type":"ContainerDied","Data":"10cb1811294b98f36229b594eb4b631d0de9e984be8bc388d78725fce94a2c5a"} Oct 07 21:17:43 crc kubenswrapper[4926]: I1007 21:17:43.999815 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f9cb5965-h6hqc"] Oct 07 21:17:44 crc kubenswrapper[4926]: I1007 21:17:44.008899 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85f9cb5965-h6hqc"] Oct 07 21:17:44 crc kubenswrapper[4926]: I1007 21:17:44.697285 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" path="/var/lib/kubelet/pods/3079e3d9-da18-49f6-b29b-275d42b4885c/volumes" Oct 07 21:17:44 crc kubenswrapper[4926]: I1007 21:17:44.723335 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" event={"ID":"f8f8c2fb-1675-4abc-8ae2-dbe68370e347","Type":"ContainerStarted","Data":"519316a5d10c2529befb2f16f36dcb4851964dcd5ac4f638bc182abc28ebf37c"} Oct 07 21:17:44 crc kubenswrapper[4926]: I1007 21:17:44.723863 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:44 crc kubenswrapper[4926]: I1007 21:17:44.752145 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" podStartSLOduration=3.75211959 podStartE2EDuration="3.75211959s" podCreationTimestamp="2025-10-07 21:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:17:44.743469938 +0000 UTC m=+1314.782051088" watchObservedRunningTime="2025-10-07 21:17:44.75211959 +0000 UTC m=+1314.790700750" Oct 07 21:17:52 crc kubenswrapper[4926]: I1007 21:17:52.007589 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c8d6588b7-27rms" Oct 07 21:17:52 crc kubenswrapper[4926]: I1007 21:17:52.073046 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867bc666f9-58g2f"] Oct 07 21:17:52 crc kubenswrapper[4926]: I1007 21:17:52.073909 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerName="dnsmasq-dns" containerID="cri-o://092780f8b17ab628d3e2f5be0f078f8e4e57bd9fc15168513a45cc8577deb6d5" gracePeriod=10 Oct 07 21:17:52 crc kubenswrapper[4926]: I1007 21:17:52.832103 4926 generic.go:334] "Generic (PLEG): container finished" podID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerID="092780f8b17ab628d3e2f5be0f078f8e4e57bd9fc15168513a45cc8577deb6d5" exitCode=0 Oct 07 21:17:52 crc kubenswrapper[4926]: I1007 21:17:52.832207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" event={"ID":"0a67502a-9033-4674-a475-3e5d3bb3d379","Type":"ContainerDied","Data":"092780f8b17ab628d3e2f5be0f078f8e4e57bd9fc15168513a45cc8577deb6d5"} Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.213024 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260095 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-svc\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260157 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-nb\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260292 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-openstack-edpm-ipam\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260375 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-config\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260461 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-sb\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260510 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-swift-storage-0\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.260586 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmc66\" (UniqueName: \"kubernetes.io/projected/0a67502a-9033-4674-a475-3e5d3bb3d379-kube-api-access-gmc66\") pod \"0a67502a-9033-4674-a475-3e5d3bb3d379\" (UID: \"0a67502a-9033-4674-a475-3e5d3bb3d379\") " Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.266236 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a67502a-9033-4674-a475-3e5d3bb3d379-kube-api-access-gmc66" (OuterVolumeSpecName: "kube-api-access-gmc66") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "kube-api-access-gmc66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.338657 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.341753 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.348429 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-config" (OuterVolumeSpecName: "config") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.351799 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.354302 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.355187 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a67502a-9033-4674-a475-3e5d3bb3d379" (UID: "0a67502a-9033-4674-a475-3e5d3bb3d379"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362723 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362743 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362753 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362762 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362772 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmc66\" (UniqueName: \"kubernetes.io/projected/0a67502a-9033-4674-a475-3e5d3bb3d379-kube-api-access-gmc66\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362783 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.362792 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a67502a-9033-4674-a475-3e5d3bb3d379-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.848826 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" event={"ID":"0a67502a-9033-4674-a475-3e5d3bb3d379","Type":"ContainerDied","Data":"034abbaa010c04aa94c4f428f67dd01ca9c3d4dceb69972fc07c29565a853f44"} Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.848916 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867bc666f9-58g2f" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.848932 4926 scope.go:117] "RemoveContainer" containerID="092780f8b17ab628d3e2f5be0f078f8e4e57bd9fc15168513a45cc8577deb6d5" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.886541 4926 scope.go:117] "RemoveContainer" containerID="126021cfb71624ff4bc09529a1b34ee8edd32c30ae300727b73d3f2d1bae712a" Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.918271 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867bc666f9-58g2f"] Oct 07 21:17:53 crc kubenswrapper[4926]: I1007 21:17:53.932219 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-867bc666f9-58g2f"] Oct 07 21:17:54 crc kubenswrapper[4926]: I1007 21:17:54.702882 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" path="/var/lib/kubelet/pods/0a67502a-9033-4674-a475-3e5d3bb3d379/volumes" Oct 07 21:17:56 crc kubenswrapper[4926]: I1007 21:17:56.904220 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5d417c9-ac1d-4b58-9506-bc6dd17a71b1" containerID="92b938d857b1e307613d07faae7fbd33fc7fe53b8d6a602a63add4986e80d112" exitCode=0 Oct 07 21:17:56 crc kubenswrapper[4926]: I1007 21:17:56.904262 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1","Type":"ContainerDied","Data":"92b938d857b1e307613d07faae7fbd33fc7fe53b8d6a602a63add4986e80d112"} Oct 07 21:17:56 crc kubenswrapper[4926]: I1007 21:17:56.908820 4926 generic.go:334] "Generic (PLEG): container finished" podID="9052310e-1ddc-4b9b-a3bc-ced3dc97ab21" containerID="6f6ae2be9b9fe803269e11f8b3855557c4dc71bcd39e32ca19cf83b1d7beb3c6" exitCode=0 Oct 07 21:17:56 crc kubenswrapper[4926]: I1007 21:17:56.908872 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21","Type":"ContainerDied","Data":"6f6ae2be9b9fe803269e11f8b3855557c4dc71bcd39e32ca19cf83b1d7beb3c6"} Oct 07 21:17:57 crc kubenswrapper[4926]: I1007 21:17:57.938681 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9052310e-1ddc-4b9b-a3bc-ced3dc97ab21","Type":"ContainerStarted","Data":"008784acadace32cce71864028b6b9fdefd20d3e8b86d5a2c12220ca185ad47c"} Oct 07 21:17:57 crc kubenswrapper[4926]: I1007 21:17:57.939521 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 21:17:57 crc kubenswrapper[4926]: I1007 21:17:57.948074 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d5d417c9-ac1d-4b58-9506-bc6dd17a71b1","Type":"ContainerStarted","Data":"c30fb47f25399be4c1e4afb9da22c73144cb9cbd0591181128c3800900828fad"} Oct 07 21:17:57 crc kubenswrapper[4926]: I1007 21:17:57.948257 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:17:57 crc kubenswrapper[4926]: I1007 21:17:57.979671 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.979646513 podStartE2EDuration="37.979646513s" podCreationTimestamp="2025-10-07 21:17:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:17:57.971715752 +0000 UTC m=+1328.010296922" watchObservedRunningTime="2025-10-07 21:17:57.979646513 +0000 UTC m=+1328.018227683" Oct 07 21:17:58 crc kubenswrapper[4926]: I1007 21:17:58.029670 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.029647256 podStartE2EDuration="37.029647256s" podCreationTimestamp="2025-10-07 21:17:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:17:57.997412039 +0000 UTC m=+1328.035993189" watchObservedRunningTime="2025-10-07 21:17:58.029647256 +0000 UTC m=+1328.068228416" Oct 07 21:18:00 crc kubenswrapper[4926]: I1007 21:18:00.860691 4926 scope.go:117] "RemoveContainer" containerID="c5849ec77af20d02832aac47aa1f49f50eb8926ee7f12ca8b44110c99e92132d" Oct 07 21:18:03 crc kubenswrapper[4926]: I1007 21:18:03.209061 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:18:03 crc kubenswrapper[4926]: I1007 21:18:03.209478 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:18:03 crc kubenswrapper[4926]: I1007 21:18:03.209540 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:18:03 crc kubenswrapper[4926]: I1007 21:18:03.210476 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65db47897203feb0e606b82feeba2549b62a2540d28ad5a38331860f14f09627"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:18:03 crc kubenswrapper[4926]: I1007 21:18:03.210552 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://65db47897203feb0e606b82feeba2549b62a2540d28ad5a38331860f14f09627" gracePeriod=600 Oct 07 21:18:04 crc kubenswrapper[4926]: I1007 21:18:04.017382 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="65db47897203feb0e606b82feeba2549b62a2540d28ad5a38331860f14f09627" exitCode=0 Oct 07 21:18:04 crc kubenswrapper[4926]: I1007 21:18:04.017477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"65db47897203feb0e606b82feeba2549b62a2540d28ad5a38331860f14f09627"} Oct 07 21:18:04 crc kubenswrapper[4926]: I1007 21:18:04.018303 4926 scope.go:117] "RemoveContainer" containerID="4c2c0ad449595a88cf92d0a873021ac44f258f8c38778755a129cc3c50f40b2f" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.031509 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752"} Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.922065 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc"] Oct 07 21:18:05 crc kubenswrapper[4926]: E1007 21:18:05.923053 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerName="dnsmasq-dns" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.923080 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerName="dnsmasq-dns" Oct 07 21:18:05 crc kubenswrapper[4926]: E1007 21:18:05.923108 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerName="init" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.923120 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerName="init" Oct 07 21:18:05 crc kubenswrapper[4926]: E1007 21:18:05.923147 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerName="dnsmasq-dns" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.923160 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerName="dnsmasq-dns" Oct 07 21:18:05 crc kubenswrapper[4926]: E1007 21:18:05.923181 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerName="init" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.923258 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerName="init" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.923664 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3079e3d9-da18-49f6-b29b-275d42b4885c" containerName="dnsmasq-dns" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.923726 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a67502a-9033-4674-a475-3e5d3bb3d379" containerName="dnsmasq-dns" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.924870 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.927971 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.928864 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.929423 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.929617 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:18:05 crc kubenswrapper[4926]: I1007 21:18:05.936530 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc"] Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.044611 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.044770 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.044803 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4bpw\" (UniqueName: \"kubernetes.io/projected/b9c1704c-58c8-407c-bafc-c23a19cf900b-kube-api-access-h4bpw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.045097 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.147497 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.147576 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.147706 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.147735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4bpw\" (UniqueName: \"kubernetes.io/projected/b9c1704c-58c8-407c-bafc-c23a19cf900b-kube-api-access-h4bpw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.153839 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.154211 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.167459 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.169732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4bpw\" (UniqueName: \"kubernetes.io/projected/b9c1704c-58c8-407c-bafc-c23a19cf900b-kube-api-access-h4bpw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.250780 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:06 crc kubenswrapper[4926]: W1007 21:18:06.897715 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9c1704c_58c8_407c_bafc_c23a19cf900b.slice/crio-157b516c3cf48581fb4e19dc240bc56acb3f228020fe4f4ee9b4b344be8b28cf WatchSource:0}: Error finding container 157b516c3cf48581fb4e19dc240bc56acb3f228020fe4f4ee9b4b344be8b28cf: Status 404 returned error can't find the container with id 157b516c3cf48581fb4e19dc240bc56acb3f228020fe4f4ee9b4b344be8b28cf Oct 07 21:18:06 crc kubenswrapper[4926]: I1007 21:18:06.903162 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc"] Oct 07 21:18:07 crc kubenswrapper[4926]: I1007 21:18:07.051760 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" event={"ID":"b9c1704c-58c8-407c-bafc-c23a19cf900b","Type":"ContainerStarted","Data":"157b516c3cf48581fb4e19dc240bc56acb3f228020fe4f4ee9b4b344be8b28cf"} Oct 07 21:18:11 crc kubenswrapper[4926]: I1007 21:18:11.049923 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9052310e-1ddc-4b9b-a3bc-ced3dc97ab21" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.228:5671: connect: connection refused" Oct 07 21:18:11 crc kubenswrapper[4926]: I1007 21:18:11.843958 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d5d417c9-ac1d-4b58-9506-bc6dd17a71b1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.229:5671: connect: connection refused" Oct 07 21:18:17 crc kubenswrapper[4926]: I1007 21:18:17.180160 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" event={"ID":"b9c1704c-58c8-407c-bafc-c23a19cf900b","Type":"ContainerStarted","Data":"c991f827bf8008ac44f1843032d507e5412668d2abaa9af20a10915319f4ce46"} Oct 07 21:18:17 crc kubenswrapper[4926]: I1007 21:18:17.215985 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" podStartSLOduration=2.661742166 podStartE2EDuration="12.215959912s" podCreationTimestamp="2025-10-07 21:18:05 +0000 UTC" firstStartedPulling="2025-10-07 21:18:06.900756421 +0000 UTC m=+1336.939337581" lastFinishedPulling="2025-10-07 21:18:16.454974167 +0000 UTC m=+1346.493555327" observedRunningTime="2025-10-07 21:18:17.207669341 +0000 UTC m=+1347.246250491" watchObservedRunningTime="2025-10-07 21:18:17.215959912 +0000 UTC m=+1347.254541082" Oct 07 21:18:21 crc kubenswrapper[4926]: I1007 21:18:21.048435 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 21:18:21 crc kubenswrapper[4926]: I1007 21:18:21.844276 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 21:18:30 crc kubenswrapper[4926]: I1007 21:18:30.308167 4926 generic.go:334] "Generic (PLEG): container finished" podID="b9c1704c-58c8-407c-bafc-c23a19cf900b" containerID="c991f827bf8008ac44f1843032d507e5412668d2abaa9af20a10915319f4ce46" exitCode=0 Oct 07 21:18:30 crc kubenswrapper[4926]: I1007 21:18:30.308229 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" event={"ID":"b9c1704c-58c8-407c-bafc-c23a19cf900b","Type":"ContainerDied","Data":"c991f827bf8008ac44f1843032d507e5412668d2abaa9af20a10915319f4ce46"} Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.848233 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.964557 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-inventory\") pod \"b9c1704c-58c8-407c-bafc-c23a19cf900b\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.964955 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4bpw\" (UniqueName: \"kubernetes.io/projected/b9c1704c-58c8-407c-bafc-c23a19cf900b-kube-api-access-h4bpw\") pod \"b9c1704c-58c8-407c-bafc-c23a19cf900b\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.965077 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-ssh-key\") pod \"b9c1704c-58c8-407c-bafc-c23a19cf900b\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.965339 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-repo-setup-combined-ca-bundle\") pod \"b9c1704c-58c8-407c-bafc-c23a19cf900b\" (UID: \"b9c1704c-58c8-407c-bafc-c23a19cf900b\") " Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.972419 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9c1704c-58c8-407c-bafc-c23a19cf900b-kube-api-access-h4bpw" (OuterVolumeSpecName: "kube-api-access-h4bpw") pod "b9c1704c-58c8-407c-bafc-c23a19cf900b" (UID: "b9c1704c-58c8-407c-bafc-c23a19cf900b"). InnerVolumeSpecName "kube-api-access-h4bpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:18:31 crc kubenswrapper[4926]: I1007 21:18:31.972663 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b9c1704c-58c8-407c-bafc-c23a19cf900b" (UID: "b9c1704c-58c8-407c-bafc-c23a19cf900b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.004184 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b9c1704c-58c8-407c-bafc-c23a19cf900b" (UID: "b9c1704c-58c8-407c-bafc-c23a19cf900b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.004735 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-inventory" (OuterVolumeSpecName: "inventory") pod "b9c1704c-58c8-407c-bafc-c23a19cf900b" (UID: "b9c1704c-58c8-407c-bafc-c23a19cf900b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.068387 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4bpw\" (UniqueName: \"kubernetes.io/projected/b9c1704c-58c8-407c-bafc-c23a19cf900b-kube-api-access-h4bpw\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.068440 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.068461 4926 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.068484 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9c1704c-58c8-407c-bafc-c23a19cf900b-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.338429 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" event={"ID":"b9c1704c-58c8-407c-bafc-c23a19cf900b","Type":"ContainerDied","Data":"157b516c3cf48581fb4e19dc240bc56acb3f228020fe4f4ee9b4b344be8b28cf"} Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.338738 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="157b516c3cf48581fb4e19dc240bc56acb3f228020fe4f4ee9b4b344be8b28cf" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.338566 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.409862 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc"] Oct 07 21:18:32 crc kubenswrapper[4926]: E1007 21:18:32.410416 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9c1704c-58c8-407c-bafc-c23a19cf900b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.410441 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9c1704c-58c8-407c-bafc-c23a19cf900b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.410653 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9c1704c-58c8-407c-bafc-c23a19cf900b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.411555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.413702 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.415057 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.415268 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.415292 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.421024 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc"] Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.476157 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.476322 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.476410 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr6rp\" (UniqueName: \"kubernetes.io/projected/b581b351-be97-4417-9c19-58e962bf821e-kube-api-access-rr6rp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.578982 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.579122 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.579250 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr6rp\" (UniqueName: \"kubernetes.io/projected/b581b351-be97-4417-9c19-58e962bf821e-kube-api-access-rr6rp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.584178 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.585154 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.604150 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr6rp\" (UniqueName: \"kubernetes.io/projected/b581b351-be97-4417-9c19-58e962bf821e-kube-api-access-rr6rp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hnftc\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:32 crc kubenswrapper[4926]: I1007 21:18:32.739006 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:33 crc kubenswrapper[4926]: W1007 21:18:33.358527 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb581b351_be97_4417_9c19_58e962bf821e.slice/crio-33ae564a3accf555d3b2f5946e3efc7ae2261dc937ce277996b2d98ddb314b0b WatchSource:0}: Error finding container 33ae564a3accf555d3b2f5946e3efc7ae2261dc937ce277996b2d98ddb314b0b: Status 404 returned error can't find the container with id 33ae564a3accf555d3b2f5946e3efc7ae2261dc937ce277996b2d98ddb314b0b Oct 07 21:18:33 crc kubenswrapper[4926]: I1007 21:18:33.361717 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc"] Oct 07 21:18:34 crc kubenswrapper[4926]: I1007 21:18:34.369632 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" event={"ID":"b581b351-be97-4417-9c19-58e962bf821e","Type":"ContainerStarted","Data":"d2e81f28d0597ae24e34500a6ac221b4b3e4f7ce925a4951439fcb8da8622781"} Oct 07 21:18:34 crc kubenswrapper[4926]: I1007 21:18:34.370477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" event={"ID":"b581b351-be97-4417-9c19-58e962bf821e","Type":"ContainerStarted","Data":"33ae564a3accf555d3b2f5946e3efc7ae2261dc937ce277996b2d98ddb314b0b"} Oct 07 21:18:34 crc kubenswrapper[4926]: I1007 21:18:34.404579 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" podStartSLOduration=1.9397761500000001 podStartE2EDuration="2.404553165s" podCreationTimestamp="2025-10-07 21:18:32 +0000 UTC" firstStartedPulling="2025-10-07 21:18:33.362718485 +0000 UTC m=+1363.401299675" lastFinishedPulling="2025-10-07 21:18:33.82749554 +0000 UTC m=+1363.866076690" observedRunningTime="2025-10-07 21:18:34.395293047 +0000 UTC m=+1364.433874267" watchObservedRunningTime="2025-10-07 21:18:34.404553165 +0000 UTC m=+1364.443134325" Oct 07 21:18:37 crc kubenswrapper[4926]: I1007 21:18:37.438073 4926 generic.go:334] "Generic (PLEG): container finished" podID="b581b351-be97-4417-9c19-58e962bf821e" containerID="d2e81f28d0597ae24e34500a6ac221b4b3e4f7ce925a4951439fcb8da8622781" exitCode=0 Oct 07 21:18:37 crc kubenswrapper[4926]: I1007 21:18:37.438141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" event={"ID":"b581b351-be97-4417-9c19-58e962bf821e","Type":"ContainerDied","Data":"d2e81f28d0597ae24e34500a6ac221b4b3e4f7ce925a4951439fcb8da8622781"} Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.001325 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.124577 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr6rp\" (UniqueName: \"kubernetes.io/projected/b581b351-be97-4417-9c19-58e962bf821e-kube-api-access-rr6rp\") pod \"b581b351-be97-4417-9c19-58e962bf821e\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.124838 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-ssh-key\") pod \"b581b351-be97-4417-9c19-58e962bf821e\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.124874 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-inventory\") pod \"b581b351-be97-4417-9c19-58e962bf821e\" (UID: \"b581b351-be97-4417-9c19-58e962bf821e\") " Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.130786 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b581b351-be97-4417-9c19-58e962bf821e-kube-api-access-rr6rp" (OuterVolumeSpecName: "kube-api-access-rr6rp") pod "b581b351-be97-4417-9c19-58e962bf821e" (UID: "b581b351-be97-4417-9c19-58e962bf821e"). InnerVolumeSpecName "kube-api-access-rr6rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.161607 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b581b351-be97-4417-9c19-58e962bf821e" (UID: "b581b351-be97-4417-9c19-58e962bf821e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.163966 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-inventory" (OuterVolumeSpecName: "inventory") pod "b581b351-be97-4417-9c19-58e962bf821e" (UID: "b581b351-be97-4417-9c19-58e962bf821e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.227660 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.227700 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b581b351-be97-4417-9c19-58e962bf821e-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.227713 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr6rp\" (UniqueName: \"kubernetes.io/projected/b581b351-be97-4417-9c19-58e962bf821e-kube-api-access-rr6rp\") on node \"crc\" DevicePath \"\"" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.461486 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" event={"ID":"b581b351-be97-4417-9c19-58e962bf821e","Type":"ContainerDied","Data":"33ae564a3accf555d3b2f5946e3efc7ae2261dc937ce277996b2d98ddb314b0b"} Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.461708 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33ae564a3accf555d3b2f5946e3efc7ae2261dc937ce277996b2d98ddb314b0b" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.461549 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hnftc" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.592731 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn"] Oct 07 21:18:39 crc kubenswrapper[4926]: E1007 21:18:39.593465 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b581b351-be97-4417-9c19-58e962bf821e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.593497 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b581b351-be97-4417-9c19-58e962bf821e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.593818 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b581b351-be97-4417-9c19-58e962bf821e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.594933 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.597658 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.597862 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.600738 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.600808 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.604623 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn"] Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.638585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.639093 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.639456 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.639534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmq6h\" (UniqueName: \"kubernetes.io/projected/9612131a-d029-4ed0-a6a0-f2562d6b34d2-kube-api-access-pmq6h\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.741607 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.741764 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.741812 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmq6h\" (UniqueName: \"kubernetes.io/projected/9612131a-d029-4ed0-a6a0-f2562d6b34d2-kube-api-access-pmq6h\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.741982 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.747769 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.748480 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.748684 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.762048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmq6h\" (UniqueName: \"kubernetes.io/projected/9612131a-d029-4ed0-a6a0-f2562d6b34d2-kube-api-access-pmq6h\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:39 crc kubenswrapper[4926]: I1007 21:18:39.928044 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:18:40 crc kubenswrapper[4926]: I1007 21:18:40.591254 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn"] Oct 07 21:18:41 crc kubenswrapper[4926]: I1007 21:18:41.494142 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" event={"ID":"9612131a-d029-4ed0-a6a0-f2562d6b34d2","Type":"ContainerStarted","Data":"429cb77504c3b8fc626fb059242f55553d821e425683398a81a53b4b0fb7259d"} Oct 07 21:18:41 crc kubenswrapper[4926]: I1007 21:18:41.494579 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" event={"ID":"9612131a-d029-4ed0-a6a0-f2562d6b34d2","Type":"ContainerStarted","Data":"1a4e9b4f1e2296991fae3a623eeb2e08a5bbf5c7414cef255de9f7b166855b4a"} Oct 07 21:18:41 crc kubenswrapper[4926]: I1007 21:18:41.530427 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" podStartSLOduration=2.008548074 podStartE2EDuration="2.530391891s" podCreationTimestamp="2025-10-07 21:18:39 +0000 UTC" firstStartedPulling="2025-10-07 21:18:40.576287671 +0000 UTC m=+1370.614868821" lastFinishedPulling="2025-10-07 21:18:41.098131458 +0000 UTC m=+1371.136712638" observedRunningTime="2025-10-07 21:18:41.52344975 +0000 UTC m=+1371.562030910" watchObservedRunningTime="2025-10-07 21:18:41.530391891 +0000 UTC m=+1371.568973111" Oct 07 21:19:01 crc kubenswrapper[4926]: I1007 21:19:01.010117 4926 scope.go:117] "RemoveContainer" containerID="e4bdac8a8da435e799222ab498f246d4a44a31aa41525ed182770d2dfb8c79e5" Oct 07 21:19:01 crc kubenswrapper[4926]: I1007 21:19:01.047556 4926 scope.go:117] "RemoveContainer" containerID="351555bd14f81eb8b0f93340e06e4d042de73cc132d3495b66c58f0be22f80ab" Oct 07 21:19:01 crc kubenswrapper[4926]: I1007 21:19:01.088680 4926 scope.go:117] "RemoveContainer" containerID="f8ff812f242d17af28a6b92666939f0ec469d943ff421daca201cc8823ac4e5e" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.158012 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fs7ln"] Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.164410 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.171518 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fs7ln"] Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.317141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-catalog-content\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.317367 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-utilities\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.317533 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4vpm\" (UniqueName: \"kubernetes.io/projected/93a41754-38f5-4464-9926-7c23eb41b9a5-kube-api-access-n4vpm\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.420378 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-catalog-content\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.420473 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-utilities\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.420546 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4vpm\" (UniqueName: \"kubernetes.io/projected/93a41754-38f5-4464-9926-7c23eb41b9a5-kube-api-access-n4vpm\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.421111 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-catalog-content\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.421255 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-utilities\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.453436 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4vpm\" (UniqueName: \"kubernetes.io/projected/93a41754-38f5-4464-9926-7c23eb41b9a5-kube-api-access-n4vpm\") pod \"certified-operators-fs7ln\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:19 crc kubenswrapper[4926]: I1007 21:19:19.500750 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:20 crc kubenswrapper[4926]: I1007 21:19:20.040613 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fs7ln"] Oct 07 21:19:21 crc kubenswrapper[4926]: I1007 21:19:21.025997 4926 generic.go:334] "Generic (PLEG): container finished" podID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerID="b58493fe1ee66bc3d884262bb923d69eae6ef293caa861ed2301177be1584480" exitCode=0 Oct 07 21:19:21 crc kubenswrapper[4926]: I1007 21:19:21.026090 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fs7ln" event={"ID":"93a41754-38f5-4464-9926-7c23eb41b9a5","Type":"ContainerDied","Data":"b58493fe1ee66bc3d884262bb923d69eae6ef293caa861ed2301177be1584480"} Oct 07 21:19:21 crc kubenswrapper[4926]: I1007 21:19:21.026510 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fs7ln" event={"ID":"93a41754-38f5-4464-9926-7c23eb41b9a5","Type":"ContainerStarted","Data":"922a30f2bec45b9a44cab5f2dc48ae347fa41e58939169803a4781aa857acba1"} Oct 07 21:19:23 crc kubenswrapper[4926]: I1007 21:19:23.056608 4926 generic.go:334] "Generic (PLEG): container finished" podID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerID="1235ac722dfe5d3a4654abf2f9b8c7cb231fe309fec7b0cf916b166d5f0175cb" exitCode=0 Oct 07 21:19:23 crc kubenswrapper[4926]: I1007 21:19:23.056719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fs7ln" event={"ID":"93a41754-38f5-4464-9926-7c23eb41b9a5","Type":"ContainerDied","Data":"1235ac722dfe5d3a4654abf2f9b8c7cb231fe309fec7b0cf916b166d5f0175cb"} Oct 07 21:19:25 crc kubenswrapper[4926]: I1007 21:19:25.076682 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fs7ln" event={"ID":"93a41754-38f5-4464-9926-7c23eb41b9a5","Type":"ContainerStarted","Data":"de6b3bb6c126c9ff9db3f345ed12b73ebfd89b1e9e8d43712519029f5300be60"} Oct 07 21:19:25 crc kubenswrapper[4926]: I1007 21:19:25.126715 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fs7ln" podStartSLOduration=3.265374887 podStartE2EDuration="6.126688149s" podCreationTimestamp="2025-10-07 21:19:19 +0000 UTC" firstStartedPulling="2025-10-07 21:19:21.029059068 +0000 UTC m=+1411.067640208" lastFinishedPulling="2025-10-07 21:19:23.89037232 +0000 UTC m=+1413.928953470" observedRunningTime="2025-10-07 21:19:25.120364876 +0000 UTC m=+1415.158946036" watchObservedRunningTime="2025-10-07 21:19:25.126688149 +0000 UTC m=+1415.165269299" Oct 07 21:19:29 crc kubenswrapper[4926]: I1007 21:19:29.501922 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:29 crc kubenswrapper[4926]: I1007 21:19:29.502513 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:29 crc kubenswrapper[4926]: I1007 21:19:29.579060 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:30 crc kubenswrapper[4926]: I1007 21:19:30.216275 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:30 crc kubenswrapper[4926]: I1007 21:19:30.308622 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fs7ln"] Oct 07 21:19:32 crc kubenswrapper[4926]: I1007 21:19:32.168856 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fs7ln" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="registry-server" containerID="cri-o://de6b3bb6c126c9ff9db3f345ed12b73ebfd89b1e9e8d43712519029f5300be60" gracePeriod=2 Oct 07 21:19:33 crc kubenswrapper[4926]: I1007 21:19:33.184602 4926 generic.go:334] "Generic (PLEG): container finished" podID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerID="de6b3bb6c126c9ff9db3f345ed12b73ebfd89b1e9e8d43712519029f5300be60" exitCode=0 Oct 07 21:19:33 crc kubenswrapper[4926]: I1007 21:19:33.184664 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fs7ln" event={"ID":"93a41754-38f5-4464-9926-7c23eb41b9a5","Type":"ContainerDied","Data":"de6b3bb6c126c9ff9db3f345ed12b73ebfd89b1e9e8d43712519029f5300be60"} Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.065147 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.141612 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-utilities\") pod \"93a41754-38f5-4464-9926-7c23eb41b9a5\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.142130 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4vpm\" (UniqueName: \"kubernetes.io/projected/93a41754-38f5-4464-9926-7c23eb41b9a5-kube-api-access-n4vpm\") pod \"93a41754-38f5-4464-9926-7c23eb41b9a5\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.142231 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-catalog-content\") pod \"93a41754-38f5-4464-9926-7c23eb41b9a5\" (UID: \"93a41754-38f5-4464-9926-7c23eb41b9a5\") " Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.143332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-utilities" (OuterVolumeSpecName: "utilities") pod "93a41754-38f5-4464-9926-7c23eb41b9a5" (UID: "93a41754-38f5-4464-9926-7c23eb41b9a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.151077 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a41754-38f5-4464-9926-7c23eb41b9a5-kube-api-access-n4vpm" (OuterVolumeSpecName: "kube-api-access-n4vpm") pod "93a41754-38f5-4464-9926-7c23eb41b9a5" (UID: "93a41754-38f5-4464-9926-7c23eb41b9a5"). InnerVolumeSpecName "kube-api-access-n4vpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.200224 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fs7ln" event={"ID":"93a41754-38f5-4464-9926-7c23eb41b9a5","Type":"ContainerDied","Data":"922a30f2bec45b9a44cab5f2dc48ae347fa41e58939169803a4781aa857acba1"} Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.200284 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fs7ln" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.200303 4926 scope.go:117] "RemoveContainer" containerID="de6b3bb6c126c9ff9db3f345ed12b73ebfd89b1e9e8d43712519029f5300be60" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.209518 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93a41754-38f5-4464-9926-7c23eb41b9a5" (UID: "93a41754-38f5-4464-9926-7c23eb41b9a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.244907 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4vpm\" (UniqueName: \"kubernetes.io/projected/93a41754-38f5-4464-9926-7c23eb41b9a5-kube-api-access-n4vpm\") on node \"crc\" DevicePath \"\"" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.244958 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.244979 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a41754-38f5-4464-9926-7c23eb41b9a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.247379 4926 scope.go:117] "RemoveContainer" containerID="1235ac722dfe5d3a4654abf2f9b8c7cb231fe309fec7b0cf916b166d5f0175cb" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.288072 4926 scope.go:117] "RemoveContainer" containerID="b58493fe1ee66bc3d884262bb923d69eae6ef293caa861ed2301177be1584480" Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.559425 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fs7ln"] Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.572841 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fs7ln"] Oct 07 21:19:34 crc kubenswrapper[4926]: I1007 21:19:34.704479 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" path="/var/lib/kubelet/pods/93a41754-38f5-4464-9926-7c23eb41b9a5/volumes" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.528372 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bfjc9"] Oct 07 21:19:39 crc kubenswrapper[4926]: E1007 21:19:39.529337 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="extract-utilities" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.529349 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="extract-utilities" Oct 07 21:19:39 crc kubenswrapper[4926]: E1007 21:19:39.529368 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="extract-content" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.529374 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="extract-content" Oct 07 21:19:39 crc kubenswrapper[4926]: E1007 21:19:39.529404 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="registry-server" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.529410 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="registry-server" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.529622 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a41754-38f5-4464-9926-7c23eb41b9a5" containerName="registry-server" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.530957 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.551675 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfjc9"] Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.670744 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4ftd\" (UniqueName: \"kubernetes.io/projected/7cab92c0-4805-4384-b48c-17bd9cd72f07-kube-api-access-q4ftd\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.670862 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-catalog-content\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.670906 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-utilities\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.772908 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4ftd\" (UniqueName: \"kubernetes.io/projected/7cab92c0-4805-4384-b48c-17bd9cd72f07-kube-api-access-q4ftd\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.773036 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-catalog-content\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.773111 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-utilities\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.773876 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-utilities\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.774234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-catalog-content\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.791883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4ftd\" (UniqueName: \"kubernetes.io/projected/7cab92c0-4805-4384-b48c-17bd9cd72f07-kube-api-access-q4ftd\") pod \"redhat-operators-bfjc9\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:39 crc kubenswrapper[4926]: I1007 21:19:39.858535 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:40 crc kubenswrapper[4926]: I1007 21:19:40.353690 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfjc9"] Oct 07 21:19:41 crc kubenswrapper[4926]: I1007 21:19:41.296138 4926 generic.go:334] "Generic (PLEG): container finished" podID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerID="88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267" exitCode=0 Oct 07 21:19:41 crc kubenswrapper[4926]: I1007 21:19:41.296247 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfjc9" event={"ID":"7cab92c0-4805-4384-b48c-17bd9cd72f07","Type":"ContainerDied","Data":"88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267"} Oct 07 21:19:41 crc kubenswrapper[4926]: I1007 21:19:41.296647 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfjc9" event={"ID":"7cab92c0-4805-4384-b48c-17bd9cd72f07","Type":"ContainerStarted","Data":"0f739b9bec2b1e35872d73b20ba4916a50178ab2a656a5411b2ee8315a717a91"} Oct 07 21:19:43 crc kubenswrapper[4926]: I1007 21:19:43.331305 4926 generic.go:334] "Generic (PLEG): container finished" podID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerID="25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408" exitCode=0 Oct 07 21:19:43 crc kubenswrapper[4926]: I1007 21:19:43.331380 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfjc9" event={"ID":"7cab92c0-4805-4384-b48c-17bd9cd72f07","Type":"ContainerDied","Data":"25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408"} Oct 07 21:19:44 crc kubenswrapper[4926]: I1007 21:19:44.353358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfjc9" event={"ID":"7cab92c0-4805-4384-b48c-17bd9cd72f07","Type":"ContainerStarted","Data":"ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659"} Oct 07 21:19:44 crc kubenswrapper[4926]: I1007 21:19:44.388920 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bfjc9" podStartSLOduration=2.89588364 podStartE2EDuration="5.38888949s" podCreationTimestamp="2025-10-07 21:19:39 +0000 UTC" firstStartedPulling="2025-10-07 21:19:41.301088341 +0000 UTC m=+1431.339669531" lastFinishedPulling="2025-10-07 21:19:43.794094181 +0000 UTC m=+1433.832675381" observedRunningTime="2025-10-07 21:19:44.375838922 +0000 UTC m=+1434.414420082" watchObservedRunningTime="2025-10-07 21:19:44.38888949 +0000 UTC m=+1434.427470640" Oct 07 21:19:49 crc kubenswrapper[4926]: I1007 21:19:49.859445 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:49 crc kubenswrapper[4926]: I1007 21:19:49.859841 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:49 crc kubenswrapper[4926]: I1007 21:19:49.965131 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:50 crc kubenswrapper[4926]: I1007 21:19:50.490962 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:50 crc kubenswrapper[4926]: I1007 21:19:50.569023 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfjc9"] Oct 07 21:19:52 crc kubenswrapper[4926]: I1007 21:19:52.453092 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bfjc9" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="registry-server" containerID="cri-o://ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659" gracePeriod=2 Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.059813 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.183551 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-utilities\") pod \"7cab92c0-4805-4384-b48c-17bd9cd72f07\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.183710 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-catalog-content\") pod \"7cab92c0-4805-4384-b48c-17bd9cd72f07\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.183885 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4ftd\" (UniqueName: \"kubernetes.io/projected/7cab92c0-4805-4384-b48c-17bd9cd72f07-kube-api-access-q4ftd\") pod \"7cab92c0-4805-4384-b48c-17bd9cd72f07\" (UID: \"7cab92c0-4805-4384-b48c-17bd9cd72f07\") " Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.184632 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-utilities" (OuterVolumeSpecName: "utilities") pod "7cab92c0-4805-4384-b48c-17bd9cd72f07" (UID: "7cab92c0-4805-4384-b48c-17bd9cd72f07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.189062 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cab92c0-4805-4384-b48c-17bd9cd72f07-kube-api-access-q4ftd" (OuterVolumeSpecName: "kube-api-access-q4ftd") pod "7cab92c0-4805-4384-b48c-17bd9cd72f07" (UID: "7cab92c0-4805-4384-b48c-17bd9cd72f07"). InnerVolumeSpecName "kube-api-access-q4ftd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.264359 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cab92c0-4805-4384-b48c-17bd9cd72f07" (UID: "7cab92c0-4805-4384-b48c-17bd9cd72f07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.286672 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4ftd\" (UniqueName: \"kubernetes.io/projected/7cab92c0-4805-4384-b48c-17bd9cd72f07-kube-api-access-q4ftd\") on node \"crc\" DevicePath \"\"" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.286701 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.286710 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cab92c0-4805-4384-b48c-17bd9cd72f07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.462660 4926 generic.go:334] "Generic (PLEG): container finished" podID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerID="ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659" exitCode=0 Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.462703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfjc9" event={"ID":"7cab92c0-4805-4384-b48c-17bd9cd72f07","Type":"ContainerDied","Data":"ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659"} Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.462712 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfjc9" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.462735 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfjc9" event={"ID":"7cab92c0-4805-4384-b48c-17bd9cd72f07","Type":"ContainerDied","Data":"0f739b9bec2b1e35872d73b20ba4916a50178ab2a656a5411b2ee8315a717a91"} Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.462755 4926 scope.go:117] "RemoveContainer" containerID="ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.486842 4926 scope.go:117] "RemoveContainer" containerID="25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.530613 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfjc9"] Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.533851 4926 scope.go:117] "RemoveContainer" containerID="88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.541309 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bfjc9"] Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.585581 4926 scope.go:117] "RemoveContainer" containerID="ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659" Oct 07 21:19:53 crc kubenswrapper[4926]: E1007 21:19:53.586016 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659\": container with ID starting with ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659 not found: ID does not exist" containerID="ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.586060 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659"} err="failed to get container status \"ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659\": rpc error: code = NotFound desc = could not find container \"ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659\": container with ID starting with ac26f2a0cbb813e9a130ff91062019ed000682c257a60da7fd91cc92a5956659 not found: ID does not exist" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.586092 4926 scope.go:117] "RemoveContainer" containerID="25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408" Oct 07 21:19:53 crc kubenswrapper[4926]: E1007 21:19:53.586461 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408\": container with ID starting with 25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408 not found: ID does not exist" containerID="25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.586488 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408"} err="failed to get container status \"25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408\": rpc error: code = NotFound desc = could not find container \"25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408\": container with ID starting with 25876ef088bfcdd0b3727c0797f89837a3e0d9f5534113ec348a4e0dd2f17408 not found: ID does not exist" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.586504 4926 scope.go:117] "RemoveContainer" containerID="88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267" Oct 07 21:19:53 crc kubenswrapper[4926]: E1007 21:19:53.586806 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267\": container with ID starting with 88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267 not found: ID does not exist" containerID="88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267" Oct 07 21:19:53 crc kubenswrapper[4926]: I1007 21:19:53.586855 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267"} err="failed to get container status \"88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267\": rpc error: code = NotFound desc = could not find container \"88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267\": container with ID starting with 88cd3a01c4582d0a10e14dc5aa97df43ea28397b994765057b66007ee04d4267 not found: ID does not exist" Oct 07 21:19:54 crc kubenswrapper[4926]: I1007 21:19:54.691948 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" path="/var/lib/kubelet/pods/7cab92c0-4805-4384-b48c-17bd9cd72f07/volumes" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.256175 4926 scope.go:117] "RemoveContainer" containerID="b5917365b23b7add7c4d569c40fd0fed548587c32043a71d30edf53797ff1b48" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.281147 4926 scope.go:117] "RemoveContainer" containerID="d7ca7a40d73b191f4315170a6719b4141987aa15743f21b39f4748f7a868f914" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.450380 4926 scope.go:117] "RemoveContainer" containerID="e8b69768c1a203c24195549018320d7b514cbec5bcdbd709dfd5fb508e3fbba4" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.499457 4926 scope.go:117] "RemoveContainer" containerID="c998b9bb106c97402a01e919798bd57bd94c1b298710141166203f882e278cb6" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.545143 4926 scope.go:117] "RemoveContainer" containerID="da72c77ea12e4db9b8a43736abc6763a069d1d927ecd4cebb44151e808ea21b2" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.589677 4926 scope.go:117] "RemoveContainer" containerID="7b99091b1c0330be8de5a61020f454ab9528baff13f2e06388e32e18ab6dc813" Oct 07 21:20:01 crc kubenswrapper[4926]: I1007 21:20:01.630104 4926 scope.go:117] "RemoveContainer" containerID="09e6b54c7a805f24299659965dc167ee53ce559d4da466c095790298d53440d5" Oct 07 21:20:33 crc kubenswrapper[4926]: I1007 21:20:33.209505 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:20:33 crc kubenswrapper[4926]: I1007 21:20:33.210041 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.842937 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5c6h"] Oct 07 21:20:59 crc kubenswrapper[4926]: E1007 21:20:59.844024 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="extract-utilities" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.844043 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="extract-utilities" Oct 07 21:20:59 crc kubenswrapper[4926]: E1007 21:20:59.844064 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="extract-content" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.844072 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="extract-content" Oct 07 21:20:59 crc kubenswrapper[4926]: E1007 21:20:59.844092 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="registry-server" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.844099 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="registry-server" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.844368 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cab92c0-4805-4384-b48c-17bd9cd72f07" containerName="registry-server" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.846148 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:20:59 crc kubenswrapper[4926]: I1007 21:20:59.873756 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5c6h"] Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.045225 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-utilities\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.045290 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsgwp\" (UniqueName: \"kubernetes.io/projected/224f5ad7-fc83-4a72-a214-2bce7ea63b28-kube-api-access-lsgwp\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.045404 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-catalog-content\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.147135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsgwp\" (UniqueName: \"kubernetes.io/projected/224f5ad7-fc83-4a72-a214-2bce7ea63b28-kube-api-access-lsgwp\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.147319 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-catalog-content\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.147537 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-utilities\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.147973 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-catalog-content\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.148034 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-utilities\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.166143 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsgwp\" (UniqueName: \"kubernetes.io/projected/224f5ad7-fc83-4a72-a214-2bce7ea63b28-kube-api-access-lsgwp\") pod \"redhat-marketplace-f5c6h\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.173643 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:00 crc kubenswrapper[4926]: I1007 21:21:00.694275 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5c6h"] Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.359494 4926 generic.go:334] "Generic (PLEG): container finished" podID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerID="85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3" exitCode=0 Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.359606 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5c6h" event={"ID":"224f5ad7-fc83-4a72-a214-2bce7ea63b28","Type":"ContainerDied","Data":"85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3"} Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.359789 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5c6h" event={"ID":"224f5ad7-fc83-4a72-a214-2bce7ea63b28","Type":"ContainerStarted","Data":"061f434562746a6bab2f6cbe1f7db3d5db97ec7001b60d87c39c40859c43f54d"} Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.364514 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.840903 4926 scope.go:117] "RemoveContainer" containerID="c97a1128966df6323a30edc576ce4e36b29dd5694c631091b9201d3843aad377" Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.879351 4926 scope.go:117] "RemoveContainer" containerID="682ec989e0923a62b8b1d4ccbe15cbce926d45bc56fc13bf4a977837b79dbec7" Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.908501 4926 scope.go:117] "RemoveContainer" containerID="19b4e58f9466c1b8227f113d5aacfb990e8ca33333ceaf1283f31330a9dd1895" Oct 07 21:21:01 crc kubenswrapper[4926]: I1007 21:21:01.936313 4926 scope.go:117] "RemoveContainer" containerID="c2978894f94e376ff151ea0b1c8a3968b11fa5960e7311b617c92b908f8e8925" Oct 07 21:21:03 crc kubenswrapper[4926]: I1007 21:21:03.209521 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:21:03 crc kubenswrapper[4926]: I1007 21:21:03.209896 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:21:03 crc kubenswrapper[4926]: I1007 21:21:03.392176 4926 generic.go:334] "Generic (PLEG): container finished" podID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerID="ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e" exitCode=0 Oct 07 21:21:03 crc kubenswrapper[4926]: I1007 21:21:03.392358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5c6h" event={"ID":"224f5ad7-fc83-4a72-a214-2bce7ea63b28","Type":"ContainerDied","Data":"ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e"} Oct 07 21:21:04 crc kubenswrapper[4926]: I1007 21:21:04.410092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5c6h" event={"ID":"224f5ad7-fc83-4a72-a214-2bce7ea63b28","Type":"ContainerStarted","Data":"892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e"} Oct 07 21:21:04 crc kubenswrapper[4926]: I1007 21:21:04.435436 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5c6h" podStartSLOduration=2.949317656 podStartE2EDuration="5.43540272s" podCreationTimestamp="2025-10-07 21:20:59 +0000 UTC" firstStartedPulling="2025-10-07 21:21:01.363893504 +0000 UTC m=+1511.402474694" lastFinishedPulling="2025-10-07 21:21:03.849978598 +0000 UTC m=+1513.888559758" observedRunningTime="2025-10-07 21:21:04.434937318 +0000 UTC m=+1514.473518478" watchObservedRunningTime="2025-10-07 21:21:04.43540272 +0000 UTC m=+1514.473983910" Oct 07 21:21:10 crc kubenswrapper[4926]: I1007 21:21:10.175040 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:10 crc kubenswrapper[4926]: I1007 21:21:10.175809 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:10 crc kubenswrapper[4926]: I1007 21:21:10.236228 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:10 crc kubenswrapper[4926]: I1007 21:21:10.553845 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:10 crc kubenswrapper[4926]: I1007 21:21:10.623123 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5c6h"] Oct 07 21:21:12 crc kubenswrapper[4926]: I1007 21:21:12.511481 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f5c6h" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="registry-server" containerID="cri-o://892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e" gracePeriod=2 Oct 07 21:21:12 crc kubenswrapper[4926]: I1007 21:21:12.974398 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.045385 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-catalog-content\") pod \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.045425 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-utilities\") pod \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.045612 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsgwp\" (UniqueName: \"kubernetes.io/projected/224f5ad7-fc83-4a72-a214-2bce7ea63b28-kube-api-access-lsgwp\") pod \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\" (UID: \"224f5ad7-fc83-4a72-a214-2bce7ea63b28\") " Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.047135 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-utilities" (OuterVolumeSpecName: "utilities") pod "224f5ad7-fc83-4a72-a214-2bce7ea63b28" (UID: "224f5ad7-fc83-4a72-a214-2bce7ea63b28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.052784 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224f5ad7-fc83-4a72-a214-2bce7ea63b28-kube-api-access-lsgwp" (OuterVolumeSpecName: "kube-api-access-lsgwp") pod "224f5ad7-fc83-4a72-a214-2bce7ea63b28" (UID: "224f5ad7-fc83-4a72-a214-2bce7ea63b28"). InnerVolumeSpecName "kube-api-access-lsgwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.061899 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "224f5ad7-fc83-4a72-a214-2bce7ea63b28" (UID: "224f5ad7-fc83-4a72-a214-2bce7ea63b28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.148905 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsgwp\" (UniqueName: \"kubernetes.io/projected/224f5ad7-fc83-4a72-a214-2bce7ea63b28-kube-api-access-lsgwp\") on node \"crc\" DevicePath \"\"" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.148947 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.148957 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224f5ad7-fc83-4a72-a214-2bce7ea63b28-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.526695 4926 generic.go:334] "Generic (PLEG): container finished" podID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerID="892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e" exitCode=0 Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.526792 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5c6h" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.526841 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5c6h" event={"ID":"224f5ad7-fc83-4a72-a214-2bce7ea63b28","Type":"ContainerDied","Data":"892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e"} Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.527320 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5c6h" event={"ID":"224f5ad7-fc83-4a72-a214-2bce7ea63b28","Type":"ContainerDied","Data":"061f434562746a6bab2f6cbe1f7db3d5db97ec7001b60d87c39c40859c43f54d"} Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.527367 4926 scope.go:117] "RemoveContainer" containerID="892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.565854 4926 scope.go:117] "RemoveContainer" containerID="ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.581130 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5c6h"] Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.598586 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5c6h"] Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.616434 4926 scope.go:117] "RemoveContainer" containerID="85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.670678 4926 scope.go:117] "RemoveContainer" containerID="892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e" Oct 07 21:21:13 crc kubenswrapper[4926]: E1007 21:21:13.671082 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e\": container with ID starting with 892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e not found: ID does not exist" containerID="892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.671120 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e"} err="failed to get container status \"892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e\": rpc error: code = NotFound desc = could not find container \"892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e\": container with ID starting with 892662cf57bd0af414bd13aead582a33ccc9f4878f2c8e3a5676d98e6ab1086e not found: ID does not exist" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.671142 4926 scope.go:117] "RemoveContainer" containerID="ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e" Oct 07 21:21:13 crc kubenswrapper[4926]: E1007 21:21:13.671502 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e\": container with ID starting with ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e not found: ID does not exist" containerID="ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.671528 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e"} err="failed to get container status \"ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e\": rpc error: code = NotFound desc = could not find container \"ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e\": container with ID starting with ec0fdbea302514addb9ceb511dee12a123c7217a00427494ed0da1e42380d09e not found: ID does not exist" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.671543 4926 scope.go:117] "RemoveContainer" containerID="85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3" Oct 07 21:21:13 crc kubenswrapper[4926]: E1007 21:21:13.671776 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3\": container with ID starting with 85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3 not found: ID does not exist" containerID="85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3" Oct 07 21:21:13 crc kubenswrapper[4926]: I1007 21:21:13.671795 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3"} err="failed to get container status \"85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3\": rpc error: code = NotFound desc = could not find container \"85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3\": container with ID starting with 85de1846877b4a24f86b44b7edcac1678d817c0785b2f99ed353d929ad00c3a3 not found: ID does not exist" Oct 07 21:21:14 crc kubenswrapper[4926]: I1007 21:21:14.699149 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" path="/var/lib/kubelet/pods/224f5ad7-fc83-4a72-a214-2bce7ea63b28/volumes" Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.209487 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.210371 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.210476 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.211892 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.212014 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" gracePeriod=600 Oct 07 21:21:33 crc kubenswrapper[4926]: E1007 21:21:33.345744 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.783876 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" exitCode=0 Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.783945 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752"} Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.784331 4926 scope.go:117] "RemoveContainer" containerID="65db47897203feb0e606b82feeba2549b62a2540d28ad5a38331860f14f09627" Oct 07 21:21:33 crc kubenswrapper[4926]: I1007 21:21:33.785133 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:21:33 crc kubenswrapper[4926]: E1007 21:21:33.785564 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:21:48 crc kubenswrapper[4926]: I1007 21:21:48.679295 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:21:48 crc kubenswrapper[4926]: E1007 21:21:48.680289 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:22:02 crc kubenswrapper[4926]: I1007 21:22:02.019132 4926 scope.go:117] "RemoveContainer" containerID="4add980846fd6d10bc5891c0d57617d4f4c9ab3f7d342059614a29700e4e918d" Oct 07 21:22:02 crc kubenswrapper[4926]: I1007 21:22:02.049917 4926 scope.go:117] "RemoveContainer" containerID="7cf9772f455f2f4530dc9e8157932056a3cd39839b9e75acea30aa87bf6afa25" Oct 07 21:22:02 crc kubenswrapper[4926]: I1007 21:22:02.688936 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:22:02 crc kubenswrapper[4926]: E1007 21:22:02.690244 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:22:13 crc kubenswrapper[4926]: I1007 21:22:13.291934 4926 generic.go:334] "Generic (PLEG): container finished" podID="9612131a-d029-4ed0-a6a0-f2562d6b34d2" containerID="429cb77504c3b8fc626fb059242f55553d821e425683398a81a53b4b0fb7259d" exitCode=0 Oct 07 21:22:13 crc kubenswrapper[4926]: I1007 21:22:13.292003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" event={"ID":"9612131a-d029-4ed0-a6a0-f2562d6b34d2","Type":"ContainerDied","Data":"429cb77504c3b8fc626fb059242f55553d821e425683398a81a53b4b0fb7259d"} Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.791312 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.950708 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmq6h\" (UniqueName: \"kubernetes.io/projected/9612131a-d029-4ed0-a6a0-f2562d6b34d2-kube-api-access-pmq6h\") pod \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.950969 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-ssh-key\") pod \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.951112 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-bootstrap-combined-ca-bundle\") pod \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.951307 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-inventory\") pod \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\" (UID: \"9612131a-d029-4ed0-a6a0-f2562d6b34d2\") " Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.961410 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9612131a-d029-4ed0-a6a0-f2562d6b34d2" (UID: "9612131a-d029-4ed0-a6a0-f2562d6b34d2"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.961827 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9612131a-d029-4ed0-a6a0-f2562d6b34d2-kube-api-access-pmq6h" (OuterVolumeSpecName: "kube-api-access-pmq6h") pod "9612131a-d029-4ed0-a6a0-f2562d6b34d2" (UID: "9612131a-d029-4ed0-a6a0-f2562d6b34d2"). InnerVolumeSpecName "kube-api-access-pmq6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:22:14 crc kubenswrapper[4926]: I1007 21:22:14.995008 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9612131a-d029-4ed0-a6a0-f2562d6b34d2" (UID: "9612131a-d029-4ed0-a6a0-f2562d6b34d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.006432 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-inventory" (OuterVolumeSpecName: "inventory") pod "9612131a-d029-4ed0-a6a0-f2562d6b34d2" (UID: "9612131a-d029-4ed0-a6a0-f2562d6b34d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.050113 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-4rl9j"] Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.054306 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.054350 4926 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.054365 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9612131a-d029-4ed0-a6a0-f2562d6b34d2-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.054380 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmq6h\" (UniqueName: \"kubernetes.io/projected/9612131a-d029-4ed0-a6a0-f2562d6b34d2-kube-api-access-pmq6h\") on node \"crc\" DevicePath \"\"" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.059991 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-4rl9j"] Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.320135 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" event={"ID":"9612131a-d029-4ed0-a6a0-f2562d6b34d2","Type":"ContainerDied","Data":"1a4e9b4f1e2296991fae3a623eeb2e08a5bbf5c7414cef255de9f7b166855b4a"} Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.320621 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a4e9b4f1e2296991fae3a623eeb2e08a5bbf5c7414cef255de9f7b166855b4a" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.320249 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.472606 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq"] Oct 07 21:22:15 crc kubenswrapper[4926]: E1007 21:22:15.473036 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="extract-utilities" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.473057 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="extract-utilities" Oct 07 21:22:15 crc kubenswrapper[4926]: E1007 21:22:15.473071 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9612131a-d029-4ed0-a6a0-f2562d6b34d2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.473079 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9612131a-d029-4ed0-a6a0-f2562d6b34d2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 21:22:15 crc kubenswrapper[4926]: E1007 21:22:15.473093 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="extract-content" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.473099 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="extract-content" Oct 07 21:22:15 crc kubenswrapper[4926]: E1007 21:22:15.473119 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="registry-server" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.473126 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="registry-server" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.473316 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9612131a-d029-4ed0-a6a0-f2562d6b34d2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.473335 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="224f5ad7-fc83-4a72-a214-2bce7ea63b28" containerName="registry-server" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.474087 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.477236 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.477463 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.477611 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.477771 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.493901 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq"] Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.570258 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snhqk\" (UniqueName: \"kubernetes.io/projected/ab42f719-a43c-425e-b0b3-aff4f812aefb-kube-api-access-snhqk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.570416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.570552 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.673650 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.673915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.674039 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snhqk\" (UniqueName: \"kubernetes.io/projected/ab42f719-a43c-425e-b0b3-aff4f812aefb-kube-api-access-snhqk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.679184 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:22:15 crc kubenswrapper[4926]: E1007 21:22:15.679523 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.683742 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.684884 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.694588 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snhqk\" (UniqueName: \"kubernetes.io/projected/ab42f719-a43c-425e-b0b3-aff4f812aefb-kube-api-access-snhqk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:15 crc kubenswrapper[4926]: I1007 21:22:15.806003 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:22:16 crc kubenswrapper[4926]: I1007 21:22:16.425892 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq"] Oct 07 21:22:16 crc kubenswrapper[4926]: I1007 21:22:16.694471 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51cda538-9dad-4aef-91ac-16af8c38e2dd" path="/var/lib/kubelet/pods/51cda538-9dad-4aef-91ac-16af8c38e2dd/volumes" Oct 07 21:22:17 crc kubenswrapper[4926]: I1007 21:22:17.351184 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" event={"ID":"ab42f719-a43c-425e-b0b3-aff4f812aefb","Type":"ContainerStarted","Data":"21a559ea3e758ffdf248bb64067de21f777fd6dbb436c1784e847505950f523b"} Oct 07 21:22:17 crc kubenswrapper[4926]: I1007 21:22:17.351580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" event={"ID":"ab42f719-a43c-425e-b0b3-aff4f812aefb","Type":"ContainerStarted","Data":"5a47c5434ec6923812f1c2f7d8eac470240b371b831504287ddace677a839266"} Oct 07 21:22:17 crc kubenswrapper[4926]: I1007 21:22:17.379604 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" podStartSLOduration=1.874298453 podStartE2EDuration="2.379579522s" podCreationTimestamp="2025-10-07 21:22:15 +0000 UTC" firstStartedPulling="2025-10-07 21:22:16.438425988 +0000 UTC m=+1586.477007148" lastFinishedPulling="2025-10-07 21:22:16.943707027 +0000 UTC m=+1586.982288217" observedRunningTime="2025-10-07 21:22:17.366327094 +0000 UTC m=+1587.404908294" watchObservedRunningTime="2025-10-07 21:22:17.379579522 +0000 UTC m=+1587.418160682" Oct 07 21:22:28 crc kubenswrapper[4926]: I1007 21:22:28.679391 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:22:28 crc kubenswrapper[4926]: E1007 21:22:28.681469 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.035525 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-dcmb9"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.050503 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-n44s7"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.062311 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-f6b6-account-create-hmpr5"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.071660 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-mm68g"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.080392 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-dcmb9"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.093688 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-mm68g"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.101892 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-f6b6-account-create-hmpr5"] Oct 07 21:22:35 crc kubenswrapper[4926]: I1007 21:22:35.109870 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-n44s7"] Oct 07 21:22:36 crc kubenswrapper[4926]: I1007 21:22:36.692409 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8c5f01-13f7-4691-b2ec-1539d9e5ee09" path="/var/lib/kubelet/pods/1d8c5f01-13f7-4691-b2ec-1539d9e5ee09/volumes" Oct 07 21:22:36 crc kubenswrapper[4926]: I1007 21:22:36.693588 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4335ef96-e8fe-4f9e-9387-821a589dd37c" path="/var/lib/kubelet/pods/4335ef96-e8fe-4f9e-9387-821a589dd37c/volumes" Oct 07 21:22:36 crc kubenswrapper[4926]: I1007 21:22:36.694169 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="599932ae-c945-4613-9312-484e6a0fce73" path="/var/lib/kubelet/pods/599932ae-c945-4613-9312-484e6a0fce73/volumes" Oct 07 21:22:36 crc kubenswrapper[4926]: I1007 21:22:36.694784 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf712ae-977e-409a-901a-13bdac028385" path="/var/lib/kubelet/pods/6bf712ae-977e-409a-901a-13bdac028385/volumes" Oct 07 21:22:43 crc kubenswrapper[4926]: I1007 21:22:43.679081 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:22:43 crc kubenswrapper[4926]: E1007 21:22:43.680520 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:22:47 crc kubenswrapper[4926]: I1007 21:22:47.042570 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-2a78-account-create-q726f"] Oct 07 21:22:47 crc kubenswrapper[4926]: I1007 21:22:47.051158 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-2a78-account-create-q726f"] Oct 07 21:22:48 crc kubenswrapper[4926]: I1007 21:22:48.694393 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcbd36c7-e9b4-47e3-82da-03fb48efad6a" path="/var/lib/kubelet/pods/bcbd36c7-e9b4-47e3-82da-03fb48efad6a/volumes" Oct 07 21:22:51 crc kubenswrapper[4926]: I1007 21:22:51.034905 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2198-account-create-zg64h"] Oct 07 21:22:51 crc kubenswrapper[4926]: I1007 21:22:51.047118 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9dc0-account-create-xbzb2"] Oct 07 21:22:51 crc kubenswrapper[4926]: I1007 21:22:51.083213 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2198-account-create-zg64h"] Oct 07 21:22:51 crc kubenswrapper[4926]: I1007 21:22:51.093044 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9dc0-account-create-xbzb2"] Oct 07 21:22:52 crc kubenswrapper[4926]: I1007 21:22:52.703001 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629a1436-9bd1-4990-8af8-7a54325da442" path="/var/lib/kubelet/pods/629a1436-9bd1-4990-8af8-7a54325da442/volumes" Oct 07 21:22:52 crc kubenswrapper[4926]: I1007 21:22:52.705804 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9d07d5c-a6d6-460d-afac-e30d5c8aff2b" path="/var/lib/kubelet/pods/e9d07d5c-a6d6-460d-afac-e30d5c8aff2b/volumes" Oct 07 21:22:56 crc kubenswrapper[4926]: I1007 21:22:56.678951 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:22:56 crc kubenswrapper[4926]: E1007 21:22:56.679697 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.760323 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6lz59"] Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.763058 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.786159 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6lz59"] Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.788833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-utilities\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.789821 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-catalog-content\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.789981 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v62wg\" (UniqueName: \"kubernetes.io/projected/b4a29bb6-5a4f-43ae-adb3-7141d078489f-kube-api-access-v62wg\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.892552 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-utilities\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.892644 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-catalog-content\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.892736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v62wg\" (UniqueName: \"kubernetes.io/projected/b4a29bb6-5a4f-43ae-adb3-7141d078489f-kube-api-access-v62wg\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.893127 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-utilities\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.893232 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-catalog-content\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:22:59 crc kubenswrapper[4926]: I1007 21:22:59.920153 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v62wg\" (UniqueName: \"kubernetes.io/projected/b4a29bb6-5a4f-43ae-adb3-7141d078489f-kube-api-access-v62wg\") pod \"community-operators-6lz59\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:00 crc kubenswrapper[4926]: I1007 21:23:00.091374 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:00 crc kubenswrapper[4926]: I1007 21:23:00.585735 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6lz59"] Oct 07 21:23:00 crc kubenswrapper[4926]: W1007 21:23:00.598312 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4a29bb6_5a4f_43ae_adb3_7141d078489f.slice/crio-e15c50f4652bf2763dd7daca3d5e61de4e5dcf980f1cd16b1403092a42c397e9 WatchSource:0}: Error finding container e15c50f4652bf2763dd7daca3d5e61de4e5dcf980f1cd16b1403092a42c397e9: Status 404 returned error can't find the container with id e15c50f4652bf2763dd7daca3d5e61de4e5dcf980f1cd16b1403092a42c397e9 Oct 07 21:23:00 crc kubenswrapper[4926]: I1007 21:23:00.868100 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerID="1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5" exitCode=0 Oct 07 21:23:00 crc kubenswrapper[4926]: I1007 21:23:00.868179 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerDied","Data":"1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5"} Oct 07 21:23:00 crc kubenswrapper[4926]: I1007 21:23:00.868556 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerStarted","Data":"e15c50f4652bf2763dd7daca3d5e61de4e5dcf980f1cd16b1403092a42c397e9"} Oct 07 21:23:01 crc kubenswrapper[4926]: I1007 21:23:01.880170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerStarted","Data":"c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b"} Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.141462 4926 scope.go:117] "RemoveContainer" containerID="eed6a6831b08b5dd4f5319631afb91ba0ca92383e5039b7b6f0969dcb249b901" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.184418 4926 scope.go:117] "RemoveContainer" containerID="82d52a86993b5f058ce88c7e5bb86b5b19c93a7762d3edca8721930668bac965" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.251745 4926 scope.go:117] "RemoveContainer" containerID="40deeb91a8cd9dfbd3aa9f20d60a2fb6536d7e4db7279aa9f59ddd4060bf8a82" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.283677 4926 scope.go:117] "RemoveContainer" containerID="077117da8b6d82eccd31763b82d008a6476b1f9534acc31072167c90fffa82c2" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.335666 4926 scope.go:117] "RemoveContainer" containerID="d3593fec014128b97a57ec9ef1c247285234c7dadcf1fc711dfffcf2448abc58" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.362357 4926 scope.go:117] "RemoveContainer" containerID="81da7eaad690d373488497e682791d3af44fe02b6e4eb4a58cfd54f0a0251897" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.413797 4926 scope.go:117] "RemoveContainer" containerID="ac4739a91966a9598e487bdbd6b342e95c5e544d2cd0937a86bc10af0f1c8343" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.448822 4926 scope.go:117] "RemoveContainer" containerID="5d7f43190c01557f3aa79392cd31a6468608f3aa9e36ade1cd77da3ef0fbbf3e" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.473156 4926 scope.go:117] "RemoveContainer" containerID="369f76d9fe89d1ab3d8b6ebd6e4dfe1154342268a288df50d47ca932ce7e3202" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.519625 4926 scope.go:117] "RemoveContainer" containerID="628d8bc50b93308b18aeb83208b9bb2b987c35195b274274dfa01631f80bd301" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.554501 4926 scope.go:117] "RemoveContainer" containerID="bbadce471ff9b97707c0094473fe87c379298d36580e666e73727c0d03562b80" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.648834 4926 scope.go:117] "RemoveContainer" containerID="87c7b36a97a6de76ce763e116a6f3dab812136a114e90a2387da5730a3c6283e" Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.895395 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerID="c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b" exitCode=0 Oct 07 21:23:02 crc kubenswrapper[4926]: I1007 21:23:02.895484 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerDied","Data":"c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b"} Oct 07 21:23:03 crc kubenswrapper[4926]: I1007 21:23:03.912458 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerStarted","Data":"615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3"} Oct 07 21:23:03 crc kubenswrapper[4926]: I1007 21:23:03.940945 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6lz59" podStartSLOduration=2.435218652 podStartE2EDuration="4.940924755s" podCreationTimestamp="2025-10-07 21:22:59 +0000 UTC" firstStartedPulling="2025-10-07 21:23:00.870289397 +0000 UTC m=+1630.908870557" lastFinishedPulling="2025-10-07 21:23:03.37599549 +0000 UTC m=+1633.414576660" observedRunningTime="2025-10-07 21:23:03.934108298 +0000 UTC m=+1633.972689458" watchObservedRunningTime="2025-10-07 21:23:03.940924755 +0000 UTC m=+1633.979505925" Oct 07 21:23:05 crc kubenswrapper[4926]: I1007 21:23:05.038486 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rghgq"] Oct 07 21:23:05 crc kubenswrapper[4926]: I1007 21:23:05.055745 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rghgq"] Oct 07 21:23:06 crc kubenswrapper[4926]: I1007 21:23:06.700024 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="decadc57-716e-46a0-a0ef-e3c964c8e5b1" path="/var/lib/kubelet/pods/decadc57-716e-46a0-a0ef-e3c964c8e5b1/volumes" Oct 07 21:23:07 crc kubenswrapper[4926]: I1007 21:23:07.679041 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:23:07 crc kubenswrapper[4926]: E1007 21:23:07.679732 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:23:09 crc kubenswrapper[4926]: I1007 21:23:09.029462 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-g49qh"] Oct 07 21:23:09 crc kubenswrapper[4926]: I1007 21:23:09.038226 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-g49qh"] Oct 07 21:23:10 crc kubenswrapper[4926]: I1007 21:23:10.092031 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:10 crc kubenswrapper[4926]: I1007 21:23:10.092116 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:10 crc kubenswrapper[4926]: I1007 21:23:10.166482 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:10 crc kubenswrapper[4926]: I1007 21:23:10.695508 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="794c22da-d189-47c3-a776-40f539508653" path="/var/lib/kubelet/pods/794c22da-d189-47c3-a776-40f539508653/volumes" Oct 07 21:23:11 crc kubenswrapper[4926]: I1007 21:23:11.077866 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:12 crc kubenswrapper[4926]: I1007 21:23:12.137520 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6lz59"] Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.010797 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6lz59" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="registry-server" containerID="cri-o://615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3" gracePeriod=2 Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.502498 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.590845 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v62wg\" (UniqueName: \"kubernetes.io/projected/b4a29bb6-5a4f-43ae-adb3-7141d078489f-kube-api-access-v62wg\") pod \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.590910 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-catalog-content\") pod \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.591104 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-utilities\") pod \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\" (UID: \"b4a29bb6-5a4f-43ae-adb3-7141d078489f\") " Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.592101 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-utilities" (OuterVolumeSpecName: "utilities") pod "b4a29bb6-5a4f-43ae-adb3-7141d078489f" (UID: "b4a29bb6-5a4f-43ae-adb3-7141d078489f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.598613 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a29bb6-5a4f-43ae-adb3-7141d078489f-kube-api-access-v62wg" (OuterVolumeSpecName: "kube-api-access-v62wg") pod "b4a29bb6-5a4f-43ae-adb3-7141d078489f" (UID: "b4a29bb6-5a4f-43ae-adb3-7141d078489f"). InnerVolumeSpecName "kube-api-access-v62wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.663781 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4a29bb6-5a4f-43ae-adb3-7141d078489f" (UID: "b4a29bb6-5a4f-43ae-adb3-7141d078489f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.693743 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v62wg\" (UniqueName: \"kubernetes.io/projected/b4a29bb6-5a4f-43ae-adb3-7141d078489f-kube-api-access-v62wg\") on node \"crc\" DevicePath \"\"" Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.693784 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:23:13 crc kubenswrapper[4926]: I1007 21:23:13.693796 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a29bb6-5a4f-43ae-adb3-7141d078489f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.024662 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerID="615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3" exitCode=0 Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.024744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerDied","Data":"615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3"} Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.024930 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lz59" event={"ID":"b4a29bb6-5a4f-43ae-adb3-7141d078489f","Type":"ContainerDied","Data":"e15c50f4652bf2763dd7daca3d5e61de4e5dcf980f1cd16b1403092a42c397e9"} Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.024950 4926 scope.go:117] "RemoveContainer" containerID="615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.024764 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lz59" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.055211 4926 scope.go:117] "RemoveContainer" containerID="c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.068773 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6lz59"] Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.084938 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6lz59"] Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.106547 4926 scope.go:117] "RemoveContainer" containerID="1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.128802 4926 scope.go:117] "RemoveContainer" containerID="615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3" Oct 07 21:23:14 crc kubenswrapper[4926]: E1007 21:23:14.129303 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3\": container with ID starting with 615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3 not found: ID does not exist" containerID="615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.129332 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3"} err="failed to get container status \"615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3\": rpc error: code = NotFound desc = could not find container \"615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3\": container with ID starting with 615b6e856a14b44ae6eaa03c211777386a944236a24b079e66e51a85d4c79ee3 not found: ID does not exist" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.129355 4926 scope.go:117] "RemoveContainer" containerID="c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b" Oct 07 21:23:14 crc kubenswrapper[4926]: E1007 21:23:14.129969 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b\": container with ID starting with c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b not found: ID does not exist" containerID="c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.130032 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b"} err="failed to get container status \"c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b\": rpc error: code = NotFound desc = could not find container \"c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b\": container with ID starting with c343d5b6071b9b3dff63319e5115b1574d72cb7275152787912895b685d8c00b not found: ID does not exist" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.130046 4926 scope.go:117] "RemoveContainer" containerID="1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5" Oct 07 21:23:14 crc kubenswrapper[4926]: E1007 21:23:14.131358 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5\": container with ID starting with 1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5 not found: ID does not exist" containerID="1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.131420 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5"} err="failed to get container status \"1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5\": rpc error: code = NotFound desc = could not find container \"1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5\": container with ID starting with 1a718dbdc38afec731318e41191a4086dc4aaa3ef2b5c018ff5ac450283c90d5 not found: ID does not exist" Oct 07 21:23:14 crc kubenswrapper[4926]: I1007 21:23:14.697958 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" path="/var/lib/kubelet/pods/b4a29bb6-5a4f-43ae-adb3-7141d078489f/volumes" Oct 07 21:23:15 crc kubenswrapper[4926]: I1007 21:23:15.064987 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-v5d87"] Oct 07 21:23:15 crc kubenswrapper[4926]: I1007 21:23:15.071516 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-v5d87"] Oct 07 21:23:16 crc kubenswrapper[4926]: I1007 21:23:16.692361 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b711cb-48ea-4199-a59e-059d35e747a6" path="/var/lib/kubelet/pods/d7b711cb-48ea-4199-a59e-059d35e747a6/volumes" Oct 07 21:23:18 crc kubenswrapper[4926]: I1007 21:23:18.028983 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-x66gt"] Oct 07 21:23:18 crc kubenswrapper[4926]: I1007 21:23:18.038299 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-x66gt"] Oct 07 21:23:18 crc kubenswrapper[4926]: I1007 21:23:18.694643 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc48d275-e265-4b1c-985c-bcee19c95284" path="/var/lib/kubelet/pods/dc48d275-e265-4b1c-985c-bcee19c95284/volumes" Oct 07 21:23:19 crc kubenswrapper[4926]: I1007 21:23:19.038082 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-9ff9-account-create-xkcnw"] Oct 07 21:23:19 crc kubenswrapper[4926]: I1007 21:23:19.046090 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-9ff9-account-create-xkcnw"] Oct 07 21:23:20 crc kubenswrapper[4926]: I1007 21:23:20.694023 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534eee94-23c7-4555-809e-9ffdb19b4e66" path="/var/lib/kubelet/pods/534eee94-23c7-4555-809e-9ffdb19b4e66/volumes" Oct 07 21:23:21 crc kubenswrapper[4926]: I1007 21:23:21.679937 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:23:21 crc kubenswrapper[4926]: E1007 21:23:21.680561 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:23:22 crc kubenswrapper[4926]: I1007 21:23:22.026240 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-vc4tq"] Oct 07 21:23:22 crc kubenswrapper[4926]: I1007 21:23:22.033965 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-vc4tq"] Oct 07 21:23:22 crc kubenswrapper[4926]: I1007 21:23:22.693476 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c175ad01-09b3-4754-960d-4bc3ef685d2e" path="/var/lib/kubelet/pods/c175ad01-09b3-4754-960d-4bc3ef685d2e/volumes" Oct 07 21:23:23 crc kubenswrapper[4926]: I1007 21:23:23.028994 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-vrdm9"] Oct 07 21:23:23 crc kubenswrapper[4926]: I1007 21:23:23.045108 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-vrdm9"] Oct 07 21:23:24 crc kubenswrapper[4926]: I1007 21:23:24.696732 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1287c912-b6cd-4814-b45e-d557e2aa4f67" path="/var/lib/kubelet/pods/1287c912-b6cd-4814-b45e-d557e2aa4f67/volumes" Oct 07 21:23:29 crc kubenswrapper[4926]: I1007 21:23:29.052621 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-bb05-account-create-6czld"] Oct 07 21:23:29 crc kubenswrapper[4926]: I1007 21:23:29.074061 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-bb05-account-create-6czld"] Oct 07 21:23:30 crc kubenswrapper[4926]: I1007 21:23:30.721040 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="806823df-1d76-4cb9-a93b-4af0cbd135e3" path="/var/lib/kubelet/pods/806823df-1d76-4cb9-a93b-4af0cbd135e3/volumes" Oct 07 21:23:32 crc kubenswrapper[4926]: I1007 21:23:32.679983 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:23:32 crc kubenswrapper[4926]: E1007 21:23:32.680614 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:23:34 crc kubenswrapper[4926]: I1007 21:23:34.047385 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-48e5-account-create-6klhk"] Oct 07 21:23:34 crc kubenswrapper[4926]: I1007 21:23:34.056256 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-48e5-account-create-6klhk"] Oct 07 21:23:34 crc kubenswrapper[4926]: I1007 21:23:34.699908 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e8a5a7f-0d0d-4213-9813-611570f3a740" path="/var/lib/kubelet/pods/6e8a5a7f-0d0d-4213-9813-611570f3a740/volumes" Oct 07 21:23:47 crc kubenswrapper[4926]: I1007 21:23:47.679307 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:23:47 crc kubenswrapper[4926]: E1007 21:23:47.680091 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:23:54 crc kubenswrapper[4926]: I1007 21:23:54.066439 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gvgtl"] Oct 07 21:23:54 crc kubenswrapper[4926]: I1007 21:23:54.079366 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gvgtl"] Oct 07 21:23:54 crc kubenswrapper[4926]: I1007 21:23:54.699189 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf7199a1-5e63-4d53-93dd-5818fb4e0454" path="/var/lib/kubelet/pods/bf7199a1-5e63-4d53-93dd-5818fb4e0454/volumes" Oct 07 21:23:57 crc kubenswrapper[4926]: I1007 21:23:57.045512 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-pvnmq"] Oct 07 21:23:57 crc kubenswrapper[4926]: I1007 21:23:57.057532 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-pvnmq"] Oct 07 21:23:58 crc kubenswrapper[4926]: I1007 21:23:58.701114 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b74a118-4677-43dd-8506-dc66795fe8a1" path="/var/lib/kubelet/pods/0b74a118-4677-43dd-8506-dc66795fe8a1/volumes" Oct 07 21:23:59 crc kubenswrapper[4926]: I1007 21:23:59.679020 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:23:59 crc kubenswrapper[4926]: E1007 21:23:59.679559 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:24:02 crc kubenswrapper[4926]: I1007 21:24:02.922035 4926 scope.go:117] "RemoveContainer" containerID="01d35b191b6388d0a11d93e61a275d4036459aae08e840f566c538d12432e275" Oct 07 21:24:02 crc kubenswrapper[4926]: I1007 21:24:02.974831 4926 scope.go:117] "RemoveContainer" containerID="6484f4ad687b7564a9b1933e1bb2a4e5569ffe21f54c4e4f7254aa5bad110678" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.054955 4926 scope.go:117] "RemoveContainer" containerID="6f408079d69f154d624584bc67a1257a946b943402c2330c77f3fcfe4add44a8" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.116453 4926 scope.go:117] "RemoveContainer" containerID="3113393e5225057a04a2b974c295774883b2e2a8c6d3ce1a4ab71497cc5d967b" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.192224 4926 scope.go:117] "RemoveContainer" containerID="7a82c2fc0970713984e825babfab5a475d63c086259872c9c94b7efb4957cda0" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.267042 4926 scope.go:117] "RemoveContainer" containerID="1872a1fee377837432fdcf378b981fb559f0eb2ac2fa9621d1cf9267d8ca2c8b" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.317695 4926 scope.go:117] "RemoveContainer" containerID="2a7495928007e508d8643e009b9be3e7f63f20d0375c98e1ad18ae4805d5e38b" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.356975 4926 scope.go:117] "RemoveContainer" containerID="518fbd721f2fdbaddfe0ff3da576094ecf27c602f6ba9ffb9d4dc2839caf860f" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.377233 4926 scope.go:117] "RemoveContainer" containerID="3f38ef18fd93b1a2ec80e15808b80341c213da5eda9506eaea8033dc03225601" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.403516 4926 scope.go:117] "RemoveContainer" containerID="ad11198ff1d392219f09037153a4f62e80cd8b0c17579626ad76b6e8b94b6d98" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.442561 4926 scope.go:117] "RemoveContainer" containerID="b74a3ca0df5a079985c18373af3386b833b6bff02f5450fcb731be1dd6c78ef6" Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.633640 4926 generic.go:334] "Generic (PLEG): container finished" podID="ab42f719-a43c-425e-b0b3-aff4f812aefb" containerID="21a559ea3e758ffdf248bb64067de21f777fd6dbb436c1784e847505950f523b" exitCode=0 Oct 07 21:24:03 crc kubenswrapper[4926]: I1007 21:24:03.633705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" event={"ID":"ab42f719-a43c-425e-b0b3-aff4f812aefb","Type":"ContainerDied","Data":"21a559ea3e758ffdf248bb64067de21f777fd6dbb436c1784e847505950f523b"} Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.277687 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.346915 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snhqk\" (UniqueName: \"kubernetes.io/projected/ab42f719-a43c-425e-b0b3-aff4f812aefb-kube-api-access-snhqk\") pod \"ab42f719-a43c-425e-b0b3-aff4f812aefb\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.347441 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-ssh-key\") pod \"ab42f719-a43c-425e-b0b3-aff4f812aefb\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.348384 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-inventory\") pod \"ab42f719-a43c-425e-b0b3-aff4f812aefb\" (UID: \"ab42f719-a43c-425e-b0b3-aff4f812aefb\") " Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.355710 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab42f719-a43c-425e-b0b3-aff4f812aefb-kube-api-access-snhqk" (OuterVolumeSpecName: "kube-api-access-snhqk") pod "ab42f719-a43c-425e-b0b3-aff4f812aefb" (UID: "ab42f719-a43c-425e-b0b3-aff4f812aefb"). InnerVolumeSpecName "kube-api-access-snhqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.389647 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab42f719-a43c-425e-b0b3-aff4f812aefb" (UID: "ab42f719-a43c-425e-b0b3-aff4f812aefb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.401670 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-inventory" (OuterVolumeSpecName: "inventory") pod "ab42f719-a43c-425e-b0b3-aff4f812aefb" (UID: "ab42f719-a43c-425e-b0b3-aff4f812aefb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.450383 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.450437 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snhqk\" (UniqueName: \"kubernetes.io/projected/ab42f719-a43c-425e-b0b3-aff4f812aefb-kube-api-access-snhqk\") on node \"crc\" DevicePath \"\"" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.450479 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab42f719-a43c-425e-b0b3-aff4f812aefb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.658440 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" event={"ID":"ab42f719-a43c-425e-b0b3-aff4f812aefb","Type":"ContainerDied","Data":"5a47c5434ec6923812f1c2f7d8eac470240b371b831504287ddace677a839266"} Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.658477 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a47c5434ec6923812f1c2f7d8eac470240b371b831504287ddace677a839266" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.658506 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.757521 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm"] Oct 07 21:24:05 crc kubenswrapper[4926]: E1007 21:24:05.757942 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="registry-server" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.757962 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="registry-server" Oct 07 21:24:05 crc kubenswrapper[4926]: E1007 21:24:05.758001 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab42f719-a43c-425e-b0b3-aff4f812aefb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.758011 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab42f719-a43c-425e-b0b3-aff4f812aefb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 07 21:24:05 crc kubenswrapper[4926]: E1007 21:24:05.758025 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="extract-utilities" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.758032 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="extract-utilities" Oct 07 21:24:05 crc kubenswrapper[4926]: E1007 21:24:05.758046 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="extract-content" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.758052 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="extract-content" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.758294 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a29bb6-5a4f-43ae-adb3-7141d078489f" containerName="registry-server" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.758318 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab42f719-a43c-425e-b0b3-aff4f812aefb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.758996 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.761761 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.762849 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.763057 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.764333 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.774345 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm"] Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.858311 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82tcx\" (UniqueName: \"kubernetes.io/projected/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-kube-api-access-82tcx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.858514 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.858596 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.960428 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82tcx\" (UniqueName: \"kubernetes.io/projected/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-kube-api-access-82tcx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.960972 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.961083 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.971141 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.972504 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:05 crc kubenswrapper[4926]: I1007 21:24:05.989281 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82tcx\" (UniqueName: \"kubernetes.io/projected/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-kube-api-access-82tcx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:06 crc kubenswrapper[4926]: I1007 21:24:06.081399 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:24:06 crc kubenswrapper[4926]: I1007 21:24:06.594619 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm"] Oct 07 21:24:06 crc kubenswrapper[4926]: I1007 21:24:06.671368 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" event={"ID":"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e","Type":"ContainerStarted","Data":"ccadce86461ae0aa44494f12b4e410ccaecd840b923161e075915c744cf646b1"} Oct 07 21:24:07 crc kubenswrapper[4926]: I1007 21:24:07.682066 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" event={"ID":"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e","Type":"ContainerStarted","Data":"18c2feebea77339efea426420c879626e393e5440d04b7038d451b407b3fc2af"} Oct 07 21:24:07 crc kubenswrapper[4926]: I1007 21:24:07.705410 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" podStartSLOduration=2.028699321 podStartE2EDuration="2.705389029s" podCreationTimestamp="2025-10-07 21:24:05 +0000 UTC" firstStartedPulling="2025-10-07 21:24:06.616491531 +0000 UTC m=+1696.655072681" lastFinishedPulling="2025-10-07 21:24:07.293181209 +0000 UTC m=+1697.331762389" observedRunningTime="2025-10-07 21:24:07.700425085 +0000 UTC m=+1697.739006275" watchObservedRunningTime="2025-10-07 21:24:07.705389029 +0000 UTC m=+1697.743970189" Oct 07 21:24:14 crc kubenswrapper[4926]: I1007 21:24:14.679931 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:24:14 crc kubenswrapper[4926]: E1007 21:24:14.681537 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:24:19 crc kubenswrapper[4926]: I1007 21:24:19.060074 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-sjjzr"] Oct 07 21:24:19 crc kubenswrapper[4926]: I1007 21:24:19.076857 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2sj9k"] Oct 07 21:24:19 crc kubenswrapper[4926]: I1007 21:24:19.084393 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2sj9k"] Oct 07 21:24:19 crc kubenswrapper[4926]: I1007 21:24:19.093137 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-sjjzr"] Oct 07 21:24:20 crc kubenswrapper[4926]: I1007 21:24:20.701784 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa8d106-5d6e-4585-b368-e54382bed918" path="/var/lib/kubelet/pods/6aa8d106-5d6e-4585-b368-e54382bed918/volumes" Oct 07 21:24:20 crc kubenswrapper[4926]: I1007 21:24:20.703512 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bab0bbe-8fa5-451f-89fe-a22426ecdfe2" path="/var/lib/kubelet/pods/6bab0bbe-8fa5-451f-89fe-a22426ecdfe2/volumes" Oct 07 21:24:28 crc kubenswrapper[4926]: I1007 21:24:28.680000 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:24:28 crc kubenswrapper[4926]: E1007 21:24:28.680984 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:24:43 crc kubenswrapper[4926]: I1007 21:24:43.681546 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:24:43 crc kubenswrapper[4926]: E1007 21:24:43.684775 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:24:47 crc kubenswrapper[4926]: I1007 21:24:47.065124 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-bszbk"] Oct 07 21:24:47 crc kubenswrapper[4926]: I1007 21:24:47.081074 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-bszbk"] Oct 07 21:24:48 crc kubenswrapper[4926]: I1007 21:24:48.703823 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="920741ff-d4f2-495b-907c-d052b32c6380" path="/var/lib/kubelet/pods/920741ff-d4f2-495b-907c-d052b32c6380/volumes" Oct 07 21:24:54 crc kubenswrapper[4926]: I1007 21:24:54.041185 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-d299q"] Oct 07 21:24:54 crc kubenswrapper[4926]: I1007 21:24:54.053516 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-d299q"] Oct 07 21:24:54 crc kubenswrapper[4926]: I1007 21:24:54.680372 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:24:54 crc kubenswrapper[4926]: E1007 21:24:54.680849 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:24:54 crc kubenswrapper[4926]: I1007 21:24:54.705573 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8039b214-d022-4525-9ae6-c7b11921aa82" path="/var/lib/kubelet/pods/8039b214-d022-4525-9ae6-c7b11921aa82/volumes" Oct 07 21:24:56 crc kubenswrapper[4926]: I1007 21:24:56.027610 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-8xkbk"] Oct 07 21:24:56 crc kubenswrapper[4926]: I1007 21:24:56.038479 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-86p2f"] Oct 07 21:24:56 crc kubenswrapper[4926]: I1007 21:24:56.063675 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-86p2f"] Oct 07 21:24:56 crc kubenswrapper[4926]: I1007 21:24:56.077562 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-8xkbk"] Oct 07 21:24:56 crc kubenswrapper[4926]: I1007 21:24:56.700675 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c212508-2303-4aac-86eb-0c5ee339e517" path="/var/lib/kubelet/pods/5c212508-2303-4aac-86eb-0c5ee339e517/volumes" Oct 07 21:24:56 crc kubenswrapper[4926]: I1007 21:24:56.702081 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db0e9ee6-03fb-45da-abda-e9af50b92dc7" path="/var/lib/kubelet/pods/db0e9ee6-03fb-45da-abda-e9af50b92dc7/volumes" Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.037033 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7d27-account-create-8482b"] Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.051663 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7d27-account-create-8482b"] Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.746509 4926 scope.go:117] "RemoveContainer" containerID="90c6e3d92954ca83fae6c6630608d9d34810f761d46a92c55da6834b453a6897" Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.805490 4926 scope.go:117] "RemoveContainer" containerID="5f1fbd0881bdd18749cb9412158fc75fa28e096f8839d0aac8c07df15707a678" Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.860961 4926 scope.go:117] "RemoveContainer" containerID="dc64d9036eb9b61d494dbf7d6dee1f340a89e9c9f72316ae6cdbbce7aa276c07" Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.938380 4926 scope.go:117] "RemoveContainer" containerID="2ef232985aea731aaf6627828bc3c183b93481e16e8852f13acd59d680db339e" Oct 07 21:25:03 crc kubenswrapper[4926]: I1007 21:25:03.965999 4926 scope.go:117] "RemoveContainer" containerID="d0e4a79821442552cfbf800b1ce377ccb5a523fb80581655d9c7a01814118561" Oct 07 21:25:04 crc kubenswrapper[4926]: I1007 21:25:04.037731 4926 scope.go:117] "RemoveContainer" containerID="029f6ea5a5085e74a95d1d3e5149a8ecc6a141bd0908624a8da52cf40222fa99" Oct 07 21:25:04 crc kubenswrapper[4926]: I1007 21:25:04.037774 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-1475-account-create-bxwm8"] Oct 07 21:25:04 crc kubenswrapper[4926]: I1007 21:25:04.046832 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-1475-account-create-bxwm8"] Oct 07 21:25:04 crc kubenswrapper[4926]: I1007 21:25:04.714052 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c7f3b3-0df4-49d5-a538-0fb9a5e10186" path="/var/lib/kubelet/pods/13c7f3b3-0df4-49d5-a538-0fb9a5e10186/volumes" Oct 07 21:25:04 crc kubenswrapper[4926]: I1007 21:25:04.715288 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7" path="/var/lib/kubelet/pods/3890b5d3-dbd2-4c4f-86c1-7c52a37a65f7/volumes" Oct 07 21:25:09 crc kubenswrapper[4926]: I1007 21:25:09.680047 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:25:09 crc kubenswrapper[4926]: E1007 21:25:09.681290 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:25:19 crc kubenswrapper[4926]: I1007 21:25:19.051474 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-587b-account-create-csjpm"] Oct 07 21:25:19 crc kubenswrapper[4926]: I1007 21:25:19.062965 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-587b-account-create-csjpm"] Oct 07 21:25:20 crc kubenswrapper[4926]: I1007 21:25:20.692997 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7375476d-b601-4885-8620-4f07b35edd55" path="/var/lib/kubelet/pods/7375476d-b601-4885-8620-4f07b35edd55/volumes" Oct 07 21:25:21 crc kubenswrapper[4926]: I1007 21:25:21.679965 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:25:21 crc kubenswrapper[4926]: E1007 21:25:21.680634 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:25:23 crc kubenswrapper[4926]: I1007 21:25:23.535244 4926 generic.go:334] "Generic (PLEG): container finished" podID="67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" containerID="18c2feebea77339efea426420c879626e393e5440d04b7038d451b407b3fc2af" exitCode=0 Oct 07 21:25:23 crc kubenswrapper[4926]: I1007 21:25:23.535339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" event={"ID":"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e","Type":"ContainerDied","Data":"18c2feebea77339efea426420c879626e393e5440d04b7038d451b407b3fc2af"} Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.145129 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.296410 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-inventory\") pod \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.297140 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-ssh-key\") pod \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.297552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82tcx\" (UniqueName: \"kubernetes.io/projected/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-kube-api-access-82tcx\") pod \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\" (UID: \"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e\") " Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.306528 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-kube-api-access-82tcx" (OuterVolumeSpecName: "kube-api-access-82tcx") pod "67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" (UID: "67f58dbd-5fcb-47a6-a904-7e3c5232dd1e"). InnerVolumeSpecName "kube-api-access-82tcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.343866 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-inventory" (OuterVolumeSpecName: "inventory") pod "67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" (UID: "67f58dbd-5fcb-47a6-a904-7e3c5232dd1e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.356658 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" (UID: "67f58dbd-5fcb-47a6-a904-7e3c5232dd1e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.400334 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82tcx\" (UniqueName: \"kubernetes.io/projected/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-kube-api-access-82tcx\") on node \"crc\" DevicePath \"\"" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.400544 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.400622 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67f58dbd-5fcb-47a6-a904-7e3c5232dd1e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.559178 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" event={"ID":"67f58dbd-5fcb-47a6-a904-7e3c5232dd1e","Type":"ContainerDied","Data":"ccadce86461ae0aa44494f12b4e410ccaecd840b923161e075915c744cf646b1"} Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.559285 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.559309 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccadce86461ae0aa44494f12b4e410ccaecd840b923161e075915c744cf646b1" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.673409 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx"] Oct 07 21:25:25 crc kubenswrapper[4926]: E1007 21:25:25.674183 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.674329 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.674665 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="67f58dbd-5fcb-47a6-a904-7e3c5232dd1e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.675609 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.682343 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.682748 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.683148 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.683430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.706809 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx"] Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.808718 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.808793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.809130 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvt5p\" (UniqueName: \"kubernetes.io/projected/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-kube-api-access-hvt5p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.911741 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.911961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvt5p\" (UniqueName: \"kubernetes.io/projected/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-kube-api-access-hvt5p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.912000 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.917834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.920963 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:25 crc kubenswrapper[4926]: I1007 21:25:25.954805 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvt5p\" (UniqueName: \"kubernetes.io/projected/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-kube-api-access-hvt5p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:26 crc kubenswrapper[4926]: I1007 21:25:26.019843 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:26 crc kubenswrapper[4926]: I1007 21:25:26.404678 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx"] Oct 07 21:25:26 crc kubenswrapper[4926]: I1007 21:25:26.571104 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" event={"ID":"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3","Type":"ContainerStarted","Data":"3839d5892bd64786faadcc475f7e8cbb65205790556986159ea85f5420778de8"} Oct 07 21:25:27 crc kubenswrapper[4926]: I1007 21:25:27.580130 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" event={"ID":"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3","Type":"ContainerStarted","Data":"896fa52465cf9e03552830dde2bb17a17412d1e5095f0134679e011cae507431"} Oct 07 21:25:32 crc kubenswrapper[4926]: I1007 21:25:32.641442 4926 generic.go:334] "Generic (PLEG): container finished" podID="a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" containerID="896fa52465cf9e03552830dde2bb17a17412d1e5095f0134679e011cae507431" exitCode=0 Oct 07 21:25:32 crc kubenswrapper[4926]: I1007 21:25:32.641549 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" event={"ID":"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3","Type":"ContainerDied","Data":"896fa52465cf9e03552830dde2bb17a17412d1e5095f0134679e011cae507431"} Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.090524 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.204310 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvt5p\" (UniqueName: \"kubernetes.io/projected/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-kube-api-access-hvt5p\") pod \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.204416 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-inventory\") pod \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.204496 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-ssh-key\") pod \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\" (UID: \"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3\") " Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.210582 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-kube-api-access-hvt5p" (OuterVolumeSpecName: "kube-api-access-hvt5p") pod "a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" (UID: "a94c25aa-d7cb-4329-8dbc-d52c8b511cf3"). InnerVolumeSpecName "kube-api-access-hvt5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.240605 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" (UID: "a94c25aa-d7cb-4329-8dbc-d52c8b511cf3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.240990 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-inventory" (OuterVolumeSpecName: "inventory") pod "a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" (UID: "a94c25aa-d7cb-4329-8dbc-d52c8b511cf3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.307224 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvt5p\" (UniqueName: \"kubernetes.io/projected/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-kube-api-access-hvt5p\") on node \"crc\" DevicePath \"\"" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.307430 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.307493 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a94c25aa-d7cb-4329-8dbc-d52c8b511cf3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.662967 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" event={"ID":"a94c25aa-d7cb-4329-8dbc-d52c8b511cf3","Type":"ContainerDied","Data":"3839d5892bd64786faadcc475f7e8cbb65205790556986159ea85f5420778de8"} Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.663454 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3839d5892bd64786faadcc475f7e8cbb65205790556986159ea85f5420778de8" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.663088 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.679314 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:25:34 crc kubenswrapper[4926]: E1007 21:25:34.680187 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.780842 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4"] Oct 07 21:25:34 crc kubenswrapper[4926]: E1007 21:25:34.781355 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.781378 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.781605 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a94c25aa-d7cb-4329-8dbc-d52c8b511cf3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.782390 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.785294 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.786146 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.786433 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.786621 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.803447 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4"] Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.816310 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.816372 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.816431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlrhs\" (UniqueName: \"kubernetes.io/projected/d64881a0-3a23-4075-94c1-46f52c09e5ef-kube-api-access-jlrhs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.918682 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.918750 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.918797 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlrhs\" (UniqueName: \"kubernetes.io/projected/d64881a0-3a23-4075-94c1-46f52c09e5ef-kube-api-access-jlrhs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.922888 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.924423 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:34 crc kubenswrapper[4926]: I1007 21:25:34.935976 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlrhs\" (UniqueName: \"kubernetes.io/projected/d64881a0-3a23-4075-94c1-46f52c09e5ef-kube-api-access-jlrhs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m9fn4\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:35 crc kubenswrapper[4926]: I1007 21:25:35.038375 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k99cm"] Oct 07 21:25:35 crc kubenswrapper[4926]: I1007 21:25:35.053644 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k99cm"] Oct 07 21:25:35 crc kubenswrapper[4926]: I1007 21:25:35.100184 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:25:35 crc kubenswrapper[4926]: W1007 21:25:35.654083 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd64881a0_3a23_4075_94c1_46f52c09e5ef.slice/crio-0f62764d615cfcabe5c917ef63e6be240f78b0576408b7364bfa0c0f368d0a2a WatchSource:0}: Error finding container 0f62764d615cfcabe5c917ef63e6be240f78b0576408b7364bfa0c0f368d0a2a: Status 404 returned error can't find the container with id 0f62764d615cfcabe5c917ef63e6be240f78b0576408b7364bfa0c0f368d0a2a Oct 07 21:25:35 crc kubenswrapper[4926]: I1007 21:25:35.662677 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4"] Oct 07 21:25:35 crc kubenswrapper[4926]: I1007 21:25:35.686364 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" event={"ID":"d64881a0-3a23-4075-94c1-46f52c09e5ef","Type":"ContainerStarted","Data":"0f62764d615cfcabe5c917ef63e6be240f78b0576408b7364bfa0c0f368d0a2a"} Oct 07 21:25:36 crc kubenswrapper[4926]: I1007 21:25:36.691432 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a4ae7fe-c26e-4dca-a071-c03a34582e53" path="/var/lib/kubelet/pods/7a4ae7fe-c26e-4dca-a071-c03a34582e53/volumes" Oct 07 21:25:36 crc kubenswrapper[4926]: I1007 21:25:36.700616 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" event={"ID":"d64881a0-3a23-4075-94c1-46f52c09e5ef","Type":"ContainerStarted","Data":"12651c3043e4c7bdff5160dd064462287fa4d268bd5d18f514ffd5d787ba176c"} Oct 07 21:25:36 crc kubenswrapper[4926]: I1007 21:25:36.732374 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" podStartSLOduration=2.263148353 podStartE2EDuration="2.732332238s" podCreationTimestamp="2025-10-07 21:25:34 +0000 UTC" firstStartedPulling="2025-10-07 21:25:35.657084126 +0000 UTC m=+1785.695665286" lastFinishedPulling="2025-10-07 21:25:36.126268021 +0000 UTC m=+1786.164849171" observedRunningTime="2025-10-07 21:25:36.72273802 +0000 UTC m=+1786.761319180" watchObservedRunningTime="2025-10-07 21:25:36.732332238 +0000 UTC m=+1786.770913438" Oct 07 21:25:49 crc kubenswrapper[4926]: I1007 21:25:49.679845 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:25:49 crc kubenswrapper[4926]: E1007 21:25:49.681329 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:26:00 crc kubenswrapper[4926]: I1007 21:26:00.066289 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zl2hw"] Oct 07 21:26:00 crc kubenswrapper[4926]: I1007 21:26:00.081492 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zl2hw"] Oct 07 21:26:00 crc kubenswrapper[4926]: I1007 21:26:00.692550 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:26:00 crc kubenswrapper[4926]: E1007 21:26:00.693679 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:26:00 crc kubenswrapper[4926]: I1007 21:26:00.695385 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a505eae2-6bc8-40cd-a133-9bd9bcb98478" path="/var/lib/kubelet/pods/a505eae2-6bc8-40cd-a133-9bd9bcb98478/volumes" Oct 07 21:26:03 crc kubenswrapper[4926]: I1007 21:26:03.060560 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbgql"] Oct 07 21:26:03 crc kubenswrapper[4926]: I1007 21:26:03.086630 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbgql"] Oct 07 21:26:04 crc kubenswrapper[4926]: I1007 21:26:04.171484 4926 scope.go:117] "RemoveContainer" containerID="b3d64569500e4791b31a167e08b98cf3b3edff5025edc1d41ca623a28eee7653" Oct 07 21:26:04 crc kubenswrapper[4926]: I1007 21:26:04.203662 4926 scope.go:117] "RemoveContainer" containerID="66d40c06f81fabf98b39eebfd6c0b56ba56d78593a1d36407ecf365eb6f9501a" Oct 07 21:26:04 crc kubenswrapper[4926]: I1007 21:26:04.275650 4926 scope.go:117] "RemoveContainer" containerID="f90f06b4e30afb44c9fdcf03e35e8aae59be7ac3b3da3be2ccf994e0944b8ccb" Oct 07 21:26:04 crc kubenswrapper[4926]: I1007 21:26:04.343663 4926 scope.go:117] "RemoveContainer" containerID="08d454e2438b30c7181ae5c3b840c1397b84367a9e4faa5c4424e26c17241dec" Oct 07 21:26:04 crc kubenswrapper[4926]: I1007 21:26:04.408136 4926 scope.go:117] "RemoveContainer" containerID="f3647224a2c7b4cf39e1d6bee3824c66704664e6234c836e868d9c7735716053" Oct 07 21:26:04 crc kubenswrapper[4926]: I1007 21:26:04.697048 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="111e3572-9e8a-4964-bc03-c5a50581e0f7" path="/var/lib/kubelet/pods/111e3572-9e8a-4964-bc03-c5a50581e0f7/volumes" Oct 07 21:26:14 crc kubenswrapper[4926]: I1007 21:26:14.679306 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:26:14 crc kubenswrapper[4926]: E1007 21:26:14.680483 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:26:16 crc kubenswrapper[4926]: I1007 21:26:16.153946 4926 generic.go:334] "Generic (PLEG): container finished" podID="d64881a0-3a23-4075-94c1-46f52c09e5ef" containerID="12651c3043e4c7bdff5160dd064462287fa4d268bd5d18f514ffd5d787ba176c" exitCode=0 Oct 07 21:26:16 crc kubenswrapper[4926]: I1007 21:26:16.154043 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" event={"ID":"d64881a0-3a23-4075-94c1-46f52c09e5ef","Type":"ContainerDied","Data":"12651c3043e4c7bdff5160dd064462287fa4d268bd5d18f514ffd5d787ba176c"} Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.698714 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.869561 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-ssh-key\") pod \"d64881a0-3a23-4075-94c1-46f52c09e5ef\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.869747 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-inventory\") pod \"d64881a0-3a23-4075-94c1-46f52c09e5ef\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.869864 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlrhs\" (UniqueName: \"kubernetes.io/projected/d64881a0-3a23-4075-94c1-46f52c09e5ef-kube-api-access-jlrhs\") pod \"d64881a0-3a23-4075-94c1-46f52c09e5ef\" (UID: \"d64881a0-3a23-4075-94c1-46f52c09e5ef\") " Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.876537 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64881a0-3a23-4075-94c1-46f52c09e5ef-kube-api-access-jlrhs" (OuterVolumeSpecName: "kube-api-access-jlrhs") pod "d64881a0-3a23-4075-94c1-46f52c09e5ef" (UID: "d64881a0-3a23-4075-94c1-46f52c09e5ef"). InnerVolumeSpecName "kube-api-access-jlrhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.907533 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d64881a0-3a23-4075-94c1-46f52c09e5ef" (UID: "d64881a0-3a23-4075-94c1-46f52c09e5ef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.920333 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-inventory" (OuterVolumeSpecName: "inventory") pod "d64881a0-3a23-4075-94c1-46f52c09e5ef" (UID: "d64881a0-3a23-4075-94c1-46f52c09e5ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.972860 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.973067 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64881a0-3a23-4075-94c1-46f52c09e5ef-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:26:17 crc kubenswrapper[4926]: I1007 21:26:17.973149 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlrhs\" (UniqueName: \"kubernetes.io/projected/d64881a0-3a23-4075-94c1-46f52c09e5ef-kube-api-access-jlrhs\") on node \"crc\" DevicePath \"\"" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.180472 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" event={"ID":"d64881a0-3a23-4075-94c1-46f52c09e5ef","Type":"ContainerDied","Data":"0f62764d615cfcabe5c917ef63e6be240f78b0576408b7364bfa0c0f368d0a2a"} Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.180526 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f62764d615cfcabe5c917ef63e6be240f78b0576408b7364bfa0c0f368d0a2a" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.180584 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m9fn4" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.290764 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb"] Oct 07 21:26:18 crc kubenswrapper[4926]: E1007 21:26:18.291245 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64881a0-3a23-4075-94c1-46f52c09e5ef" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.291263 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64881a0-3a23-4075-94c1-46f52c09e5ef" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.291447 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64881a0-3a23-4075-94c1-46f52c09e5ef" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.292114 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.295530 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.295554 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.295607 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.297426 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.304259 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb"] Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.381167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.381664 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.381795 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28dxn\" (UniqueName: \"kubernetes.io/projected/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-kube-api-access-28dxn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.484309 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.484365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.484484 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28dxn\" (UniqueName: \"kubernetes.io/projected/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-kube-api-access-28dxn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.501069 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.501363 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.511954 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28dxn\" (UniqueName: \"kubernetes.io/projected/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-kube-api-access-28dxn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rllwb\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:18 crc kubenswrapper[4926]: I1007 21:26:18.622613 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:26:19 crc kubenswrapper[4926]: I1007 21:26:19.208325 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:26:19 crc kubenswrapper[4926]: I1007 21:26:19.217991 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb"] Oct 07 21:26:20 crc kubenswrapper[4926]: I1007 21:26:20.202011 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" event={"ID":"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47","Type":"ContainerStarted","Data":"38a6bc9e1768527f3090ff837592ac952ba68c1e5e4877b1cea377d1afe029b1"} Oct 07 21:26:20 crc kubenswrapper[4926]: I1007 21:26:20.202749 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" event={"ID":"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47","Type":"ContainerStarted","Data":"620688b367c1ee49874e5f46a2fdee30ef05489214f2521e01dd150952a82e67"} Oct 07 21:26:20 crc kubenswrapper[4926]: I1007 21:26:20.221394 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" podStartSLOduration=1.687315946 podStartE2EDuration="2.221369628s" podCreationTimestamp="2025-10-07 21:26:18 +0000 UTC" firstStartedPulling="2025-10-07 21:26:19.20806967 +0000 UTC m=+1829.246650820" lastFinishedPulling="2025-10-07 21:26:19.742123362 +0000 UTC m=+1829.780704502" observedRunningTime="2025-10-07 21:26:20.220946736 +0000 UTC m=+1830.259527926" watchObservedRunningTime="2025-10-07 21:26:20.221369628 +0000 UTC m=+1830.259950788" Oct 07 21:26:28 crc kubenswrapper[4926]: I1007 21:26:28.679420 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:26:28 crc kubenswrapper[4926]: E1007 21:26:28.680347 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:26:40 crc kubenswrapper[4926]: I1007 21:26:40.695824 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:26:41 crc kubenswrapper[4926]: I1007 21:26:41.434351 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"b0fadebd37b106a55391d1cc2cbe36eb8d7a6c92c1c552b732e0960ab706c9f6"} Oct 07 21:26:43 crc kubenswrapper[4926]: I1007 21:26:43.064171 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-l8s8j"] Oct 07 21:26:43 crc kubenswrapper[4926]: I1007 21:26:43.076879 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-l8s8j"] Oct 07 21:26:44 crc kubenswrapper[4926]: I1007 21:26:44.695419 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6aa769-924b-42ea-b9f2-76f2d77abe3a" path="/var/lib/kubelet/pods/4c6aa769-924b-42ea-b9f2-76f2d77abe3a/volumes" Oct 07 21:27:04 crc kubenswrapper[4926]: I1007 21:27:04.554671 4926 scope.go:117] "RemoveContainer" containerID="379872a104dbc4211471dfe06bd17267e6e5c5c3cbb195aaa20b1e77d143e971" Oct 07 21:27:04 crc kubenswrapper[4926]: I1007 21:27:04.624161 4926 scope.go:117] "RemoveContainer" containerID="332c5791f8740dd0c6f97e584f92d04ad29eb986458d02666f1d29ffbd04864e" Oct 07 21:27:18 crc kubenswrapper[4926]: I1007 21:27:18.887088 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" containerID="38a6bc9e1768527f3090ff837592ac952ba68c1e5e4877b1cea377d1afe029b1" exitCode=2 Oct 07 21:27:18 crc kubenswrapper[4926]: I1007 21:27:18.887245 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" event={"ID":"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47","Type":"ContainerDied","Data":"38a6bc9e1768527f3090ff837592ac952ba68c1e5e4877b1cea377d1afe029b1"} Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.368173 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.456922 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-inventory\") pod \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.457004 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28dxn\" (UniqueName: \"kubernetes.io/projected/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-kube-api-access-28dxn\") pod \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.457077 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-ssh-key\") pod \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\" (UID: \"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47\") " Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.463594 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-kube-api-access-28dxn" (OuterVolumeSpecName: "kube-api-access-28dxn") pod "2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" (UID: "2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47"). InnerVolumeSpecName "kube-api-access-28dxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.485910 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-inventory" (OuterVolumeSpecName: "inventory") pod "2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" (UID: "2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.506883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" (UID: "2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.560282 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.560334 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28dxn\" (UniqueName: \"kubernetes.io/projected/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-kube-api-access-28dxn\") on node \"crc\" DevicePath \"\"" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.560356 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.918346 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" event={"ID":"2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47","Type":"ContainerDied","Data":"620688b367c1ee49874e5f46a2fdee30ef05489214f2521e01dd150952a82e67"} Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.918743 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="620688b367c1ee49874e5f46a2fdee30ef05489214f2521e01dd150952a82e67" Oct 07 21:27:20 crc kubenswrapper[4926]: I1007 21:27:20.918833 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rllwb" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.072526 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27"] Oct 07 21:27:28 crc kubenswrapper[4926]: E1007 21:27:28.073949 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.073972 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.074266 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.075316 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.077431 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.078589 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.080560 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.080570 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.083259 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27"] Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.232806 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.232928 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzqhq\" (UniqueName: \"kubernetes.io/projected/3f3a65d1-ea19-4587-a483-ba006c9d4660-kube-api-access-gzqhq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.233083 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.334911 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.334964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzqhq\" (UniqueName: \"kubernetes.io/projected/3f3a65d1-ea19-4587-a483-ba006c9d4660-kube-api-access-gzqhq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.335042 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.344179 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.350288 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.360321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzqhq\" (UniqueName: \"kubernetes.io/projected/3f3a65d1-ea19-4587-a483-ba006c9d4660-kube-api-access-gzqhq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhm27\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.393186 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:27:28 crc kubenswrapper[4926]: I1007 21:27:28.932535 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27"] Oct 07 21:27:29 crc kubenswrapper[4926]: I1007 21:27:29.018771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" event={"ID":"3f3a65d1-ea19-4587-a483-ba006c9d4660","Type":"ContainerStarted","Data":"8b98e0fcfd7d0eb6f0afa93f0027c8b0a4e24c85013d5f40b3b636aea45acea6"} Oct 07 21:27:30 crc kubenswrapper[4926]: I1007 21:27:30.028182 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" event={"ID":"3f3a65d1-ea19-4587-a483-ba006c9d4660","Type":"ContainerStarted","Data":"38bfdbba0213e5fd1754b5003108e534f8fc2a229d1fee45937070ca52e8e225"} Oct 07 21:27:30 crc kubenswrapper[4926]: I1007 21:27:30.069247 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" podStartSLOduration=1.648173557 podStartE2EDuration="2.069214547s" podCreationTimestamp="2025-10-07 21:27:28 +0000 UTC" firstStartedPulling="2025-10-07 21:27:28.944123033 +0000 UTC m=+1898.982704223" lastFinishedPulling="2025-10-07 21:27:29.365164033 +0000 UTC m=+1899.403745213" observedRunningTime="2025-10-07 21:27:30.064526221 +0000 UTC m=+1900.103107371" watchObservedRunningTime="2025-10-07 21:27:30.069214547 +0000 UTC m=+1900.107795697" Oct 07 21:28:24 crc kubenswrapper[4926]: I1007 21:28:24.694669 4926 generic.go:334] "Generic (PLEG): container finished" podID="3f3a65d1-ea19-4587-a483-ba006c9d4660" containerID="38bfdbba0213e5fd1754b5003108e534f8fc2a229d1fee45937070ca52e8e225" exitCode=0 Oct 07 21:28:24 crc kubenswrapper[4926]: I1007 21:28:24.698594 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" event={"ID":"3f3a65d1-ea19-4587-a483-ba006c9d4660","Type":"ContainerDied","Data":"38bfdbba0213e5fd1754b5003108e534f8fc2a229d1fee45937070ca52e8e225"} Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.179260 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.364745 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-inventory\") pod \"3f3a65d1-ea19-4587-a483-ba006c9d4660\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.364904 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzqhq\" (UniqueName: \"kubernetes.io/projected/3f3a65d1-ea19-4587-a483-ba006c9d4660-kube-api-access-gzqhq\") pod \"3f3a65d1-ea19-4587-a483-ba006c9d4660\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.365052 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-ssh-key\") pod \"3f3a65d1-ea19-4587-a483-ba006c9d4660\" (UID: \"3f3a65d1-ea19-4587-a483-ba006c9d4660\") " Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.371130 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f3a65d1-ea19-4587-a483-ba006c9d4660-kube-api-access-gzqhq" (OuterVolumeSpecName: "kube-api-access-gzqhq") pod "3f3a65d1-ea19-4587-a483-ba006c9d4660" (UID: "3f3a65d1-ea19-4587-a483-ba006c9d4660"). InnerVolumeSpecName "kube-api-access-gzqhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.396454 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-inventory" (OuterVolumeSpecName: "inventory") pod "3f3a65d1-ea19-4587-a483-ba006c9d4660" (UID: "3f3a65d1-ea19-4587-a483-ba006c9d4660"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.410211 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f3a65d1-ea19-4587-a483-ba006c9d4660" (UID: "3f3a65d1-ea19-4587-a483-ba006c9d4660"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.467803 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.467849 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f3a65d1-ea19-4587-a483-ba006c9d4660-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.467864 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzqhq\" (UniqueName: \"kubernetes.io/projected/3f3a65d1-ea19-4587-a483-ba006c9d4660-kube-api-access-gzqhq\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.723808 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" event={"ID":"3f3a65d1-ea19-4587-a483-ba006c9d4660","Type":"ContainerDied","Data":"8b98e0fcfd7d0eb6f0afa93f0027c8b0a4e24c85013d5f40b3b636aea45acea6"} Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.723848 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b98e0fcfd7d0eb6f0afa93f0027c8b0a4e24c85013d5f40b3b636aea45acea6" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.723863 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhm27" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.833062 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5rwsd"] Oct 07 21:28:26 crc kubenswrapper[4926]: E1007 21:28:26.833551 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f3a65d1-ea19-4587-a483-ba006c9d4660" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.833576 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f3a65d1-ea19-4587-a483-ba006c9d4660" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.833860 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f3a65d1-ea19-4587-a483-ba006c9d4660" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.834750 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.837450 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.837649 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.837855 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.837999 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.842878 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5rwsd"] Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.977391 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.977708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr7xz\" (UniqueName: \"kubernetes.io/projected/69fe71d5-112a-45aa-a043-8629073ec0c2-kube-api-access-mr7xz\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:26 crc kubenswrapper[4926]: I1007 21:28:26.978026 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.079925 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.080057 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr7xz\" (UniqueName: \"kubernetes.io/projected/69fe71d5-112a-45aa-a043-8629073ec0c2-kube-api-access-mr7xz\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.080184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.085224 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.090729 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.108347 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr7xz\" (UniqueName: \"kubernetes.io/projected/69fe71d5-112a-45aa-a043-8629073ec0c2-kube-api-access-mr7xz\") pod \"ssh-known-hosts-edpm-deployment-5rwsd\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.155369 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.715590 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5rwsd"] Oct 07 21:28:27 crc kubenswrapper[4926]: I1007 21:28:27.733761 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" event={"ID":"69fe71d5-112a-45aa-a043-8629073ec0c2","Type":"ContainerStarted","Data":"5f1b6b8e6029046fe7cb21532e2bdcf85777a48cf7a256ecbbff9fae011aee06"} Oct 07 21:28:28 crc kubenswrapper[4926]: I1007 21:28:28.744513 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" event={"ID":"69fe71d5-112a-45aa-a043-8629073ec0c2","Type":"ContainerStarted","Data":"789efccf4c08efa0e62831df36c19d16c297bb221071f7ab6a6bac501f6afcb9"} Oct 07 21:28:28 crc kubenswrapper[4926]: I1007 21:28:28.777612 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" podStartSLOduration=2.324206455 podStartE2EDuration="2.777559761s" podCreationTimestamp="2025-10-07 21:28:26 +0000 UTC" firstStartedPulling="2025-10-07 21:28:27.722327429 +0000 UTC m=+1957.760908579" lastFinishedPulling="2025-10-07 21:28:28.175680735 +0000 UTC m=+1958.214261885" observedRunningTime="2025-10-07 21:28:28.765322707 +0000 UTC m=+1958.803903907" watchObservedRunningTime="2025-10-07 21:28:28.777559761 +0000 UTC m=+1958.816140931" Oct 07 21:28:36 crc kubenswrapper[4926]: I1007 21:28:36.834138 4926 generic.go:334] "Generic (PLEG): container finished" podID="69fe71d5-112a-45aa-a043-8629073ec0c2" containerID="789efccf4c08efa0e62831df36c19d16c297bb221071f7ab6a6bac501f6afcb9" exitCode=0 Oct 07 21:28:36 crc kubenswrapper[4926]: I1007 21:28:36.834263 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" event={"ID":"69fe71d5-112a-45aa-a043-8629073ec0c2","Type":"ContainerDied","Data":"789efccf4c08efa0e62831df36c19d16c297bb221071f7ab6a6bac501f6afcb9"} Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.286867 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.426269 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-inventory-0\") pod \"69fe71d5-112a-45aa-a043-8629073ec0c2\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.426392 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-ssh-key-openstack-edpm-ipam\") pod \"69fe71d5-112a-45aa-a043-8629073ec0c2\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.426493 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr7xz\" (UniqueName: \"kubernetes.io/projected/69fe71d5-112a-45aa-a043-8629073ec0c2-kube-api-access-mr7xz\") pod \"69fe71d5-112a-45aa-a043-8629073ec0c2\" (UID: \"69fe71d5-112a-45aa-a043-8629073ec0c2\") " Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.441495 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69fe71d5-112a-45aa-a043-8629073ec0c2-kube-api-access-mr7xz" (OuterVolumeSpecName: "kube-api-access-mr7xz") pod "69fe71d5-112a-45aa-a043-8629073ec0c2" (UID: "69fe71d5-112a-45aa-a043-8629073ec0c2"). InnerVolumeSpecName "kube-api-access-mr7xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.459252 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "69fe71d5-112a-45aa-a043-8629073ec0c2" (UID: "69fe71d5-112a-45aa-a043-8629073ec0c2"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.465342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "69fe71d5-112a-45aa-a043-8629073ec0c2" (UID: "69fe71d5-112a-45aa-a043-8629073ec0c2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.528902 4926 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.529271 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/69fe71d5-112a-45aa-a043-8629073ec0c2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.529298 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr7xz\" (UniqueName: \"kubernetes.io/projected/69fe71d5-112a-45aa-a043-8629073ec0c2-kube-api-access-mr7xz\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.867383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" event={"ID":"69fe71d5-112a-45aa-a043-8629073ec0c2","Type":"ContainerDied","Data":"5f1b6b8e6029046fe7cb21532e2bdcf85777a48cf7a256ecbbff9fae011aee06"} Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.867430 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f1b6b8e6029046fe7cb21532e2bdcf85777a48cf7a256ecbbff9fae011aee06" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.867544 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5rwsd" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.948407 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm"] Oct 07 21:28:38 crc kubenswrapper[4926]: E1007 21:28:38.949061 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fe71d5-112a-45aa-a043-8629073ec0c2" containerName="ssh-known-hosts-edpm-deployment" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.949166 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fe71d5-112a-45aa-a043-8629073ec0c2" containerName="ssh-known-hosts-edpm-deployment" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.949452 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fe71d5-112a-45aa-a043-8629073ec0c2" containerName="ssh-known-hosts-edpm-deployment" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.950171 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.952642 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.952894 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.953135 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.952905 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:28:38 crc kubenswrapper[4926]: I1007 21:28:38.957694 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm"] Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.040097 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.040523 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zht6\" (UniqueName: \"kubernetes.io/projected/67068a37-7d4a-47ca-9971-89e1e7a49a25-kube-api-access-7zht6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.040618 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.143139 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zht6\" (UniqueName: \"kubernetes.io/projected/67068a37-7d4a-47ca-9971-89e1e7a49a25-kube-api-access-7zht6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.143654 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.144100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.147914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.152620 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.165977 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zht6\" (UniqueName: \"kubernetes.io/projected/67068a37-7d4a-47ca-9971-89e1e7a49a25-kube-api-access-7zht6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kbwzm\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.270549 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:39 crc kubenswrapper[4926]: I1007 21:28:39.914220 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm"] Oct 07 21:28:40 crc kubenswrapper[4926]: I1007 21:28:40.924885 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" event={"ID":"67068a37-7d4a-47ca-9971-89e1e7a49a25","Type":"ContainerStarted","Data":"f4bbbba1017d8f7d43a5331a9bc76d76c6f2df35fc8f28e5c2c84448d079a872"} Oct 07 21:28:40 crc kubenswrapper[4926]: I1007 21:28:40.925703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" event={"ID":"67068a37-7d4a-47ca-9971-89e1e7a49a25","Type":"ContainerStarted","Data":"d051823274245f05a721fa5970845bd796a077441e80789ab4fcf129aafb8702"} Oct 07 21:28:40 crc kubenswrapper[4926]: I1007 21:28:40.952575 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" podStartSLOduration=2.364635674 podStartE2EDuration="2.952548815s" podCreationTimestamp="2025-10-07 21:28:38 +0000 UTC" firstStartedPulling="2025-10-07 21:28:39.920495366 +0000 UTC m=+1969.959076556" lastFinishedPulling="2025-10-07 21:28:40.508408547 +0000 UTC m=+1970.546989697" observedRunningTime="2025-10-07 21:28:40.947605752 +0000 UTC m=+1970.986186962" watchObservedRunningTime="2025-10-07 21:28:40.952548815 +0000 UTC m=+1970.991129975" Oct 07 21:28:51 crc kubenswrapper[4926]: I1007 21:28:51.072383 4926 generic.go:334] "Generic (PLEG): container finished" podID="67068a37-7d4a-47ca-9971-89e1e7a49a25" containerID="f4bbbba1017d8f7d43a5331a9bc76d76c6f2df35fc8f28e5c2c84448d079a872" exitCode=0 Oct 07 21:28:51 crc kubenswrapper[4926]: I1007 21:28:51.073084 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" event={"ID":"67068a37-7d4a-47ca-9971-89e1e7a49a25","Type":"ContainerDied","Data":"f4bbbba1017d8f7d43a5331a9bc76d76c6f2df35fc8f28e5c2c84448d079a872"} Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.547766 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.714854 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zht6\" (UniqueName: \"kubernetes.io/projected/67068a37-7d4a-47ca-9971-89e1e7a49a25-kube-api-access-7zht6\") pod \"67068a37-7d4a-47ca-9971-89e1e7a49a25\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.715135 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-inventory\") pod \"67068a37-7d4a-47ca-9971-89e1e7a49a25\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.715389 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-ssh-key\") pod \"67068a37-7d4a-47ca-9971-89e1e7a49a25\" (UID: \"67068a37-7d4a-47ca-9971-89e1e7a49a25\") " Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.722475 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67068a37-7d4a-47ca-9971-89e1e7a49a25-kube-api-access-7zht6" (OuterVolumeSpecName: "kube-api-access-7zht6") pod "67068a37-7d4a-47ca-9971-89e1e7a49a25" (UID: "67068a37-7d4a-47ca-9971-89e1e7a49a25"). InnerVolumeSpecName "kube-api-access-7zht6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.759447 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-inventory" (OuterVolumeSpecName: "inventory") pod "67068a37-7d4a-47ca-9971-89e1e7a49a25" (UID: "67068a37-7d4a-47ca-9971-89e1e7a49a25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.771126 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "67068a37-7d4a-47ca-9971-89e1e7a49a25" (UID: "67068a37-7d4a-47ca-9971-89e1e7a49a25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.817982 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zht6\" (UniqueName: \"kubernetes.io/projected/67068a37-7d4a-47ca-9971-89e1e7a49a25-kube-api-access-7zht6\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.818030 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:52 crc kubenswrapper[4926]: I1007 21:28:52.818046 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67068a37-7d4a-47ca-9971-89e1e7a49a25-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.096442 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" event={"ID":"67068a37-7d4a-47ca-9971-89e1e7a49a25","Type":"ContainerDied","Data":"d051823274245f05a721fa5970845bd796a077441e80789ab4fcf129aafb8702"} Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.096508 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d051823274245f05a721fa5970845bd796a077441e80789ab4fcf129aafb8702" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.096549 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kbwzm" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.197425 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd"] Oct 07 21:28:53 crc kubenswrapper[4926]: E1007 21:28:53.197816 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67068a37-7d4a-47ca-9971-89e1e7a49a25" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.197834 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="67068a37-7d4a-47ca-9971-89e1e7a49a25" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.198029 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="67068a37-7d4a-47ca-9971-89e1e7a49a25" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.198738 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.202124 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.202405 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.202597 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.203989 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.214224 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd"] Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.228131 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.228258 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bd9p\" (UniqueName: \"kubernetes.io/projected/28a80874-caa4-48eb-9cbc-999e2db615fb-kube-api-access-7bd9p\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.228409 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.329751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.329803 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.329898 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bd9p\" (UniqueName: \"kubernetes.io/projected/28a80874-caa4-48eb-9cbc-999e2db615fb-kube-api-access-7bd9p\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.335242 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.336880 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.352314 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bd9p\" (UniqueName: \"kubernetes.io/projected/28a80874-caa4-48eb-9cbc-999e2db615fb-kube-api-access-7bd9p\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:53 crc kubenswrapper[4926]: I1007 21:28:53.521141 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:28:54 crc kubenswrapper[4926]: W1007 21:28:54.091493 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28a80874_caa4_48eb_9cbc_999e2db615fb.slice/crio-160a95fc1dfd6987e56aff5e1267b23589c595006d2d5d0c09f5ad53397cb31e WatchSource:0}: Error finding container 160a95fc1dfd6987e56aff5e1267b23589c595006d2d5d0c09f5ad53397cb31e: Status 404 returned error can't find the container with id 160a95fc1dfd6987e56aff5e1267b23589c595006d2d5d0c09f5ad53397cb31e Oct 07 21:28:54 crc kubenswrapper[4926]: I1007 21:28:54.097165 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd"] Oct 07 21:28:54 crc kubenswrapper[4926]: I1007 21:28:54.109767 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" event={"ID":"28a80874-caa4-48eb-9cbc-999e2db615fb","Type":"ContainerStarted","Data":"160a95fc1dfd6987e56aff5e1267b23589c595006d2d5d0c09f5ad53397cb31e"} Oct 07 21:28:55 crc kubenswrapper[4926]: I1007 21:28:55.123795 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" event={"ID":"28a80874-caa4-48eb-9cbc-999e2db615fb","Type":"ContainerStarted","Data":"1e073e8d286f92c374d902fb7b54b52d7b936b13f025448e68343b6a28536518"} Oct 07 21:28:55 crc kubenswrapper[4926]: I1007 21:28:55.158659 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" podStartSLOduration=1.68081432 podStartE2EDuration="2.158639304s" podCreationTimestamp="2025-10-07 21:28:53 +0000 UTC" firstStartedPulling="2025-10-07 21:28:54.098232044 +0000 UTC m=+1984.136813204" lastFinishedPulling="2025-10-07 21:28:54.576057038 +0000 UTC m=+1984.614638188" observedRunningTime="2025-10-07 21:28:55.144362341 +0000 UTC m=+1985.182943501" watchObservedRunningTime="2025-10-07 21:28:55.158639304 +0000 UTC m=+1985.197220464" Oct 07 21:29:03 crc kubenswrapper[4926]: I1007 21:29:03.209830 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:29:03 crc kubenswrapper[4926]: I1007 21:29:03.211053 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:29:06 crc kubenswrapper[4926]: I1007 21:29:06.273495 4926 generic.go:334] "Generic (PLEG): container finished" podID="28a80874-caa4-48eb-9cbc-999e2db615fb" containerID="1e073e8d286f92c374d902fb7b54b52d7b936b13f025448e68343b6a28536518" exitCode=0 Oct 07 21:29:06 crc kubenswrapper[4926]: I1007 21:29:06.273590 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" event={"ID":"28a80874-caa4-48eb-9cbc-999e2db615fb","Type":"ContainerDied","Data":"1e073e8d286f92c374d902fb7b54b52d7b936b13f025448e68343b6a28536518"} Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.743105 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.871364 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-inventory\") pod \"28a80874-caa4-48eb-9cbc-999e2db615fb\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.871917 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-ssh-key\") pod \"28a80874-caa4-48eb-9cbc-999e2db615fb\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.872647 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bd9p\" (UniqueName: \"kubernetes.io/projected/28a80874-caa4-48eb-9cbc-999e2db615fb-kube-api-access-7bd9p\") pod \"28a80874-caa4-48eb-9cbc-999e2db615fb\" (UID: \"28a80874-caa4-48eb-9cbc-999e2db615fb\") " Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.877993 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a80874-caa4-48eb-9cbc-999e2db615fb-kube-api-access-7bd9p" (OuterVolumeSpecName: "kube-api-access-7bd9p") pod "28a80874-caa4-48eb-9cbc-999e2db615fb" (UID: "28a80874-caa4-48eb-9cbc-999e2db615fb"). InnerVolumeSpecName "kube-api-access-7bd9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.911525 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "28a80874-caa4-48eb-9cbc-999e2db615fb" (UID: "28a80874-caa4-48eb-9cbc-999e2db615fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.915515 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-inventory" (OuterVolumeSpecName: "inventory") pod "28a80874-caa4-48eb-9cbc-999e2db615fb" (UID: "28a80874-caa4-48eb-9cbc-999e2db615fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.976684 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bd9p\" (UniqueName: \"kubernetes.io/projected/28a80874-caa4-48eb-9cbc-999e2db615fb-kube-api-access-7bd9p\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.976727 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:07 crc kubenswrapper[4926]: I1007 21:29:07.976738 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28a80874-caa4-48eb-9cbc-999e2db615fb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.299177 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" event={"ID":"28a80874-caa4-48eb-9cbc-999e2db615fb","Type":"ContainerDied","Data":"160a95fc1dfd6987e56aff5e1267b23589c595006d2d5d0c09f5ad53397cb31e"} Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.299275 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.299284 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="160a95fc1dfd6987e56aff5e1267b23589c595006d2d5d0c09f5ad53397cb31e" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.449220 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf"] Oct 07 21:29:08 crc kubenswrapper[4926]: E1007 21:29:08.449685 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a80874-caa4-48eb-9cbc-999e2db615fb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.449709 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a80874-caa4-48eb-9cbc-999e2db615fb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.449996 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a80874-caa4-48eb-9cbc-999e2db615fb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.451045 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.456248 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.456460 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.457125 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.457336 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.457180 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.458038 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.458177 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.458401 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.465251 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf"] Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591558 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591641 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtscv\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-kube-api-access-qtscv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591663 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591695 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591853 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.591936 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592118 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592172 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592279 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592372 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592507 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592571 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.592803 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694440 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694469 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694505 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtscv\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-kube-api-access-qtscv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694534 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694561 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694626 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694718 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694744 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694867 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694924 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.694954 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.698964 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.699323 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.700282 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.701980 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.702029 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.702708 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.703046 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.705675 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.705887 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.706410 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.706527 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.706688 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.707690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.717489 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtscv\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-kube-api-access-qtscv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:08 crc kubenswrapper[4926]: I1007 21:29:08.794688 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:09 crc kubenswrapper[4926]: I1007 21:29:09.425439 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf"] Oct 07 21:29:10 crc kubenswrapper[4926]: I1007 21:29:10.326888 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" event={"ID":"52cd3a5e-b86a-49c9-84fd-387676280d79","Type":"ContainerStarted","Data":"07fcf0033968ecd3b308934f8885533b8ef377a3711f059a9d103766cd8c0a77"} Oct 07 21:29:10 crc kubenswrapper[4926]: I1007 21:29:10.327112 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" event={"ID":"52cd3a5e-b86a-49c9-84fd-387676280d79","Type":"ContainerStarted","Data":"fe5297136869365af3bff5b770af836d1231d947622c17f0037c5622e2455f77"} Oct 07 21:29:10 crc kubenswrapper[4926]: I1007 21:29:10.380360 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" podStartSLOduration=1.923876433 podStartE2EDuration="2.380333809s" podCreationTimestamp="2025-10-07 21:29:08 +0000 UTC" firstStartedPulling="2025-10-07 21:29:09.408341337 +0000 UTC m=+1999.446922497" lastFinishedPulling="2025-10-07 21:29:09.864798683 +0000 UTC m=+1999.903379873" observedRunningTime="2025-10-07 21:29:10.365619383 +0000 UTC m=+2000.404200563" watchObservedRunningTime="2025-10-07 21:29:10.380333809 +0000 UTC m=+2000.418914979" Oct 07 21:29:33 crc kubenswrapper[4926]: I1007 21:29:33.209143 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:29:33 crc kubenswrapper[4926]: I1007 21:29:33.209984 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.324821 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jhcl4"] Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.330873 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.351862 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhcl4"] Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.456505 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6zwj\" (UniqueName: \"kubernetes.io/projected/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-kube-api-access-m6zwj\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.457042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-utilities\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.457107 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-catalog-content\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.558876 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-utilities\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.558942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-catalog-content\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.559030 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6zwj\" (UniqueName: \"kubernetes.io/projected/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-kube-api-access-m6zwj\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.559442 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-utilities\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.559546 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-catalog-content\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.578741 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6zwj\" (UniqueName: \"kubernetes.io/projected/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-kube-api-access-m6zwj\") pod \"redhat-operators-jhcl4\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:46 crc kubenswrapper[4926]: I1007 21:29:46.663905 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:47 crc kubenswrapper[4926]: I1007 21:29:47.173751 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhcl4"] Oct 07 21:29:47 crc kubenswrapper[4926]: I1007 21:29:47.815404 4926 generic.go:334] "Generic (PLEG): container finished" podID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerID="fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef" exitCode=0 Oct 07 21:29:47 crc kubenswrapper[4926]: I1007 21:29:47.815517 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerDied","Data":"fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef"} Oct 07 21:29:47 crc kubenswrapper[4926]: I1007 21:29:47.815811 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerStarted","Data":"037abddcab111c65a8cd9199fe89c0db8b3dbfd8ca3ec79d30dcf154323409f3"} Oct 07 21:29:49 crc kubenswrapper[4926]: I1007 21:29:49.843359 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerStarted","Data":"ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6"} Oct 07 21:29:51 crc kubenswrapper[4926]: I1007 21:29:51.881810 4926 generic.go:334] "Generic (PLEG): container finished" podID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerID="ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6" exitCode=0 Oct 07 21:29:51 crc kubenswrapper[4926]: I1007 21:29:51.881896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerDied","Data":"ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6"} Oct 07 21:29:52 crc kubenswrapper[4926]: I1007 21:29:52.895142 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerStarted","Data":"9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec"} Oct 07 21:29:52 crc kubenswrapper[4926]: I1007 21:29:52.926861 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jhcl4" podStartSLOduration=2.336641967 podStartE2EDuration="6.926831392s" podCreationTimestamp="2025-10-07 21:29:46 +0000 UTC" firstStartedPulling="2025-10-07 21:29:47.818886496 +0000 UTC m=+2037.857467636" lastFinishedPulling="2025-10-07 21:29:52.409075891 +0000 UTC m=+2042.447657061" observedRunningTime="2025-10-07 21:29:52.916547183 +0000 UTC m=+2042.955128363" watchObservedRunningTime="2025-10-07 21:29:52.926831392 +0000 UTC m=+2042.965412582" Oct 07 21:29:56 crc kubenswrapper[4926]: I1007 21:29:56.664109 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:56 crc kubenswrapper[4926]: I1007 21:29:56.664805 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:29:57 crc kubenswrapper[4926]: I1007 21:29:57.748078 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jhcl4" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="registry-server" probeResult="failure" output=< Oct 07 21:29:57 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Oct 07 21:29:57 crc kubenswrapper[4926]: > Oct 07 21:29:57 crc kubenswrapper[4926]: I1007 21:29:57.978529 4926 generic.go:334] "Generic (PLEG): container finished" podID="52cd3a5e-b86a-49c9-84fd-387676280d79" containerID="07fcf0033968ecd3b308934f8885533b8ef377a3711f059a9d103766cd8c0a77" exitCode=0 Oct 07 21:29:57 crc kubenswrapper[4926]: I1007 21:29:57.978589 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" event={"ID":"52cd3a5e-b86a-49c9-84fd-387676280d79","Type":"ContainerDied","Data":"07fcf0033968ecd3b308934f8885533b8ef377a3711f059a9d103766cd8c0a77"} Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.487019 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571249 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-ovn-default-certs-0\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571323 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ovn-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571357 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-nova-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571404 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571450 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-bootstrap-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571528 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-inventory\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571600 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ssh-key\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571629 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-libvirt-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571715 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-neutron-metadata-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571771 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571805 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtscv\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-kube-api-access-qtscv\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-telemetry-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-repo-setup-combined-ca-bundle\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.571965 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"52cd3a5e-b86a-49c9-84fd-387676280d79\" (UID: \"52cd3a5e-b86a-49c9-84fd-387676280d79\") " Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.579901 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.579939 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.579960 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.580259 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.580610 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.580887 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.580927 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-kube-api-access-qtscv" (OuterVolumeSpecName: "kube-api-access-qtscv") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "kube-api-access-qtscv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.582088 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.582458 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.584596 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.585454 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.587501 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.607181 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-inventory" (OuterVolumeSpecName: "inventory") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.615039 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "52cd3a5e-b86a-49c9-84fd-387676280d79" (UID: "52cd3a5e-b86a-49c9-84fd-387676280d79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.673942 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.673971 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtscv\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-kube-api-access-qtscv\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.673982 4926 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.673990 4926 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.673999 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674009 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674018 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674027 4926 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674035 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/52cd3a5e-b86a-49c9-84fd-387676280d79-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674043 4926 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674051 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674058 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674066 4926 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:29:59 crc kubenswrapper[4926]: I1007 21:29:59.674074 4926 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd3a5e-b86a-49c9-84fd-387676280d79-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.004075 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" event={"ID":"52cd3a5e-b86a-49c9-84fd-387676280d79","Type":"ContainerDied","Data":"fe5297136869365af3bff5b770af836d1231d947622c17f0037c5622e2455f77"} Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.004355 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe5297136869365af3bff5b770af836d1231d947622c17f0037c5622e2455f77" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.004183 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.126466 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z"] Oct 07 21:30:00 crc kubenswrapper[4926]: E1007 21:30:00.127265 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52cd3a5e-b86a-49c9-84fd-387676280d79" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.127394 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="52cd3a5e-b86a-49c9-84fd-387676280d79" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.127795 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="52cd3a5e-b86a-49c9-84fd-387676280d79" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.128762 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.131642 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.132235 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.132429 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.132617 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.140398 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.148632 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z"] Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.223981 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs"] Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.225155 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.227312 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.229762 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.249010 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs"] Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.290059 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.290386 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdtc8\" (UniqueName: \"kubernetes.io/projected/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-kube-api-access-fdtc8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.290622 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.290810 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.290942 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.392844 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c513138c-31c6-4b48-94c1-71660c5af953-secret-volume\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.393143 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.393296 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c513138c-31c6-4b48-94c1-71660c5af953-config-volume\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.393505 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.393668 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.393833 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdtc8\" (UniqueName: \"kubernetes.io/projected/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-kube-api-access-fdtc8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.393962 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.394075 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7jd7\" (UniqueName: \"kubernetes.io/projected/c513138c-31c6-4b48-94c1-71660c5af953-kube-api-access-r7jd7\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.395406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.400419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.401778 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.413163 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.418509 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdtc8\" (UniqueName: \"kubernetes.io/projected/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-kube-api-access-fdtc8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwx7z\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.459142 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.496110 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c513138c-31c6-4b48-94c1-71660c5af953-secret-volume\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.496269 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c513138c-31c6-4b48-94c1-71660c5af953-config-volume\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.496464 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7jd7\" (UniqueName: \"kubernetes.io/projected/c513138c-31c6-4b48-94c1-71660c5af953-kube-api-access-r7jd7\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.498244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c513138c-31c6-4b48-94c1-71660c5af953-config-volume\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.502925 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c513138c-31c6-4b48-94c1-71660c5af953-secret-volume\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.519874 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7jd7\" (UniqueName: \"kubernetes.io/projected/c513138c-31c6-4b48-94c1-71660c5af953-kube-api-access-r7jd7\") pod \"collect-profiles-29331210-bphvs\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:00 crc kubenswrapper[4926]: I1007 21:30:00.547268 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:01 crc kubenswrapper[4926]: I1007 21:30:01.077863 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs"] Oct 07 21:30:01 crc kubenswrapper[4926]: I1007 21:30:01.104340 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z"] Oct 07 21:30:02 crc kubenswrapper[4926]: I1007 21:30:02.038898 4926 generic.go:334] "Generic (PLEG): container finished" podID="c513138c-31c6-4b48-94c1-71660c5af953" containerID="d6644ae4256546b50892aae953071a6ef593bb9df8619e18928756fdef1d6d06" exitCode=0 Oct 07 21:30:02 crc kubenswrapper[4926]: I1007 21:30:02.039043 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" event={"ID":"c513138c-31c6-4b48-94c1-71660c5af953","Type":"ContainerDied","Data":"d6644ae4256546b50892aae953071a6ef593bb9df8619e18928756fdef1d6d06"} Oct 07 21:30:02 crc kubenswrapper[4926]: I1007 21:30:02.040883 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" event={"ID":"c513138c-31c6-4b48-94c1-71660c5af953","Type":"ContainerStarted","Data":"fa1e026c2a6e528a7c66b913073cf25cf1f7c3d2575830013fe340ec3017ccf6"} Oct 07 21:30:02 crc kubenswrapper[4926]: I1007 21:30:02.042079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" event={"ID":"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12","Type":"ContainerStarted","Data":"9473bf20c5e79d9370068903ba58784402ed923d92868dcf6425b6af17d6a7ad"} Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.075710 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" event={"ID":"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12","Type":"ContainerStarted","Data":"78ee9c2597196361ef2eecfa28478f0c4e4d15f045b13f3d22f17d62ff77c42d"} Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.108185 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" podStartSLOduration=2.173355632 podStartE2EDuration="3.108161188s" podCreationTimestamp="2025-10-07 21:30:00 +0000 UTC" firstStartedPulling="2025-10-07 21:30:01.123143434 +0000 UTC m=+2051.161724604" lastFinishedPulling="2025-10-07 21:30:02.05794901 +0000 UTC m=+2052.096530160" observedRunningTime="2025-10-07 21:30:03.105759048 +0000 UTC m=+2053.144340328" watchObservedRunningTime="2025-10-07 21:30:03.108161188 +0000 UTC m=+2053.146742348" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.209676 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.209737 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.209796 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.210569 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0fadebd37b106a55391d1cc2cbe36eb8d7a6c92c1c552b732e0960ab706c9f6"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.210636 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://b0fadebd37b106a55391d1cc2cbe36eb8d7a6c92c1c552b732e0960ab706c9f6" gracePeriod=600 Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.514848 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.675280 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c513138c-31c6-4b48-94c1-71660c5af953-secret-volume\") pod \"c513138c-31c6-4b48-94c1-71660c5af953\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.675731 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7jd7\" (UniqueName: \"kubernetes.io/projected/c513138c-31c6-4b48-94c1-71660c5af953-kube-api-access-r7jd7\") pod \"c513138c-31c6-4b48-94c1-71660c5af953\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.675846 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c513138c-31c6-4b48-94c1-71660c5af953-config-volume\") pod \"c513138c-31c6-4b48-94c1-71660c5af953\" (UID: \"c513138c-31c6-4b48-94c1-71660c5af953\") " Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.676859 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c513138c-31c6-4b48-94c1-71660c5af953-config-volume" (OuterVolumeSpecName: "config-volume") pod "c513138c-31c6-4b48-94c1-71660c5af953" (UID: "c513138c-31c6-4b48-94c1-71660c5af953"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.680958 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c513138c-31c6-4b48-94c1-71660c5af953-kube-api-access-r7jd7" (OuterVolumeSpecName: "kube-api-access-r7jd7") pod "c513138c-31c6-4b48-94c1-71660c5af953" (UID: "c513138c-31c6-4b48-94c1-71660c5af953"). InnerVolumeSpecName "kube-api-access-r7jd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.682158 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c513138c-31c6-4b48-94c1-71660c5af953-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c513138c-31c6-4b48-94c1-71660c5af953" (UID: "c513138c-31c6-4b48-94c1-71660c5af953"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.778660 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c513138c-31c6-4b48-94c1-71660c5af953-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.778698 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7jd7\" (UniqueName: \"kubernetes.io/projected/c513138c-31c6-4b48-94c1-71660c5af953-kube-api-access-r7jd7\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:03 crc kubenswrapper[4926]: I1007 21:30:03.778709 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c513138c-31c6-4b48-94c1-71660c5af953-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.088612 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="b0fadebd37b106a55391d1cc2cbe36eb8d7a6c92c1c552b732e0960ab706c9f6" exitCode=0 Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.088670 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"b0fadebd37b106a55391d1cc2cbe36eb8d7a6c92c1c552b732e0960ab706c9f6"} Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.089158 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8"} Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.089185 4926 scope.go:117] "RemoveContainer" containerID="7130e88f6a7e835fccd4482eea86d6e7b3e8a2094ecbdd66e74d6575d86d1752" Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.093055 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" event={"ID":"c513138c-31c6-4b48-94c1-71660c5af953","Type":"ContainerDied","Data":"fa1e026c2a6e528a7c66b913073cf25cf1f7c3d2575830013fe340ec3017ccf6"} Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.093107 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa1e026c2a6e528a7c66b913073cf25cf1f7c3d2575830013fe340ec3017ccf6" Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.093066 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs" Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.618070 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2"] Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.629140 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331165-4pss2"] Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.697431 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="738dcd07-f77b-4737-8195-07e674320384" path="/var/lib/kubelet/pods/738dcd07-f77b-4737-8195-07e674320384/volumes" Oct 07 21:30:04 crc kubenswrapper[4926]: I1007 21:30:04.846168 4926 scope.go:117] "RemoveContainer" containerID="ba557e2ac9cdc96297d3c0fc084ef707e8ec2d9d96224aea6096ee318ca53b9b" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.153411 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hlzfr"] Oct 07 21:30:05 crc kubenswrapper[4926]: E1007 21:30:05.164601 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c513138c-31c6-4b48-94c1-71660c5af953" containerName="collect-profiles" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.164649 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c513138c-31c6-4b48-94c1-71660c5af953" containerName="collect-profiles" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.165033 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c513138c-31c6-4b48-94c1-71660c5af953" containerName="collect-profiles" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.167472 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.176056 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hlzfr"] Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.309300 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-catalog-content\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.309393 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4zcp\" (UniqueName: \"kubernetes.io/projected/1a204473-1a29-4367-af55-8acb6cf4aa17-kube-api-access-l4zcp\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.309419 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-utilities\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.411176 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-catalog-content\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.411290 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4zcp\" (UniqueName: \"kubernetes.io/projected/1a204473-1a29-4367-af55-8acb6cf4aa17-kube-api-access-l4zcp\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.411320 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-utilities\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.411890 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-utilities\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.412089 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-catalog-content\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.440572 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4zcp\" (UniqueName: \"kubernetes.io/projected/1a204473-1a29-4367-af55-8acb6cf4aa17-kube-api-access-l4zcp\") pod \"certified-operators-hlzfr\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:05 crc kubenswrapper[4926]: I1007 21:30:05.492754 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:06 crc kubenswrapper[4926]: I1007 21:30:06.012806 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hlzfr"] Oct 07 21:30:06 crc kubenswrapper[4926]: W1007 21:30:06.018930 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a204473_1a29_4367_af55_8acb6cf4aa17.slice/crio-40782470f4499c2cc52a751d5de57b547c8883db9174c5612ebc6597d7fb5814 WatchSource:0}: Error finding container 40782470f4499c2cc52a751d5de57b547c8883db9174c5612ebc6597d7fb5814: Status 404 returned error can't find the container with id 40782470f4499c2cc52a751d5de57b547c8883db9174c5612ebc6597d7fb5814 Oct 07 21:30:06 crc kubenswrapper[4926]: I1007 21:30:06.121280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerStarted","Data":"40782470f4499c2cc52a751d5de57b547c8883db9174c5612ebc6597d7fb5814"} Oct 07 21:30:06 crc kubenswrapper[4926]: I1007 21:30:06.730821 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:30:06 crc kubenswrapper[4926]: I1007 21:30:06.814028 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:30:07 crc kubenswrapper[4926]: I1007 21:30:07.137500 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerID="cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6" exitCode=0 Oct 07 21:30:07 crc kubenswrapper[4926]: I1007 21:30:07.137595 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerDied","Data":"cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6"} Oct 07 21:30:08 crc kubenswrapper[4926]: I1007 21:30:08.149673 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerStarted","Data":"892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1"} Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.121395 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhcl4"] Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.121937 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jhcl4" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="registry-server" containerID="cri-o://9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec" gracePeriod=2 Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.173094 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerID="892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1" exitCode=0 Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.173157 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerDied","Data":"892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1"} Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.656687 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.824407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-utilities\") pod \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.824590 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6zwj\" (UniqueName: \"kubernetes.io/projected/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-kube-api-access-m6zwj\") pod \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.824726 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-catalog-content\") pod \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\" (UID: \"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2\") " Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.825664 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-utilities" (OuterVolumeSpecName: "utilities") pod "db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" (UID: "db4d34cb-9dea-4f1f-9949-de74aa1d2ab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.844385 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-kube-api-access-m6zwj" (OuterVolumeSpecName: "kube-api-access-m6zwj") pod "db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" (UID: "db4d34cb-9dea-4f1f-9949-de74aa1d2ab2"). InnerVolumeSpecName "kube-api-access-m6zwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.925454 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" (UID: "db4d34cb-9dea-4f1f-9949-de74aa1d2ab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.927001 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6zwj\" (UniqueName: \"kubernetes.io/projected/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-kube-api-access-m6zwj\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.927038 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:09 crc kubenswrapper[4926]: I1007 21:30:09.927048 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.194749 4926 generic.go:334] "Generic (PLEG): container finished" podID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerID="9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec" exitCode=0 Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.194836 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhcl4" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.194869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerDied","Data":"9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec"} Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.194938 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhcl4" event={"ID":"db4d34cb-9dea-4f1f-9949-de74aa1d2ab2","Type":"ContainerDied","Data":"037abddcab111c65a8cd9199fe89c0db8b3dbfd8ca3ec79d30dcf154323409f3"} Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.194970 4926 scope.go:117] "RemoveContainer" containerID="9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.203741 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerStarted","Data":"5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51"} Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.229679 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hlzfr" podStartSLOduration=2.7464310579999998 podStartE2EDuration="5.229656807s" podCreationTimestamp="2025-10-07 21:30:05 +0000 UTC" firstStartedPulling="2025-10-07 21:30:07.139807537 +0000 UTC m=+2057.178388697" lastFinishedPulling="2025-10-07 21:30:09.623033296 +0000 UTC m=+2059.661614446" observedRunningTime="2025-10-07 21:30:10.227138984 +0000 UTC m=+2060.265720164" watchObservedRunningTime="2025-10-07 21:30:10.229656807 +0000 UTC m=+2060.268237967" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.242603 4926 scope.go:117] "RemoveContainer" containerID="ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.255187 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhcl4"] Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.272759 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jhcl4"] Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.274327 4926 scope.go:117] "RemoveContainer" containerID="fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.324439 4926 scope.go:117] "RemoveContainer" containerID="9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec" Oct 07 21:30:10 crc kubenswrapper[4926]: E1007 21:30:10.324848 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec\": container with ID starting with 9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec not found: ID does not exist" containerID="9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.324878 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec"} err="failed to get container status \"9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec\": rpc error: code = NotFound desc = could not find container \"9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec\": container with ID starting with 9b013579b64c5ac3eb386700a19d125bcc4962242d4e36bf03f82c9577ab1fec not found: ID does not exist" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.324899 4926 scope.go:117] "RemoveContainer" containerID="ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6" Oct 07 21:30:10 crc kubenswrapper[4926]: E1007 21:30:10.325550 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6\": container with ID starting with ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6 not found: ID does not exist" containerID="ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.325584 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6"} err="failed to get container status \"ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6\": rpc error: code = NotFound desc = could not find container \"ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6\": container with ID starting with ce7d2b9c99ddd459418a1a15dbbdf9f50e5114ef1f3edbbee8984bf08ea83db6 not found: ID does not exist" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.325605 4926 scope.go:117] "RemoveContainer" containerID="fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef" Oct 07 21:30:10 crc kubenswrapper[4926]: E1007 21:30:10.325869 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef\": container with ID starting with fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef not found: ID does not exist" containerID="fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.325892 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef"} err="failed to get container status \"fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef\": rpc error: code = NotFound desc = could not find container \"fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef\": container with ID starting with fe4601df3042da3a6e217b7283a1304b85fa0fea92c75cb5e86332f003463eef not found: ID does not exist" Oct 07 21:30:10 crc kubenswrapper[4926]: I1007 21:30:10.700703 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" path="/var/lib/kubelet/pods/db4d34cb-9dea-4f1f-9949-de74aa1d2ab2/volumes" Oct 07 21:30:15 crc kubenswrapper[4926]: I1007 21:30:15.493316 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:15 crc kubenswrapper[4926]: I1007 21:30:15.493974 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:15 crc kubenswrapper[4926]: I1007 21:30:15.559585 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:16 crc kubenswrapper[4926]: I1007 21:30:16.343887 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:16 crc kubenswrapper[4926]: I1007 21:30:16.837154 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hlzfr"] Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.295853 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hlzfr" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="registry-server" containerID="cri-o://5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51" gracePeriod=2 Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.810415 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.927864 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-utilities\") pod \"1a204473-1a29-4367-af55-8acb6cf4aa17\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.927927 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4zcp\" (UniqueName: \"kubernetes.io/projected/1a204473-1a29-4367-af55-8acb6cf4aa17-kube-api-access-l4zcp\") pod \"1a204473-1a29-4367-af55-8acb6cf4aa17\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.928012 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-catalog-content\") pod \"1a204473-1a29-4367-af55-8acb6cf4aa17\" (UID: \"1a204473-1a29-4367-af55-8acb6cf4aa17\") " Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.929386 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-utilities" (OuterVolumeSpecName: "utilities") pod "1a204473-1a29-4367-af55-8acb6cf4aa17" (UID: "1a204473-1a29-4367-af55-8acb6cf4aa17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.935836 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a204473-1a29-4367-af55-8acb6cf4aa17-kube-api-access-l4zcp" (OuterVolumeSpecName: "kube-api-access-l4zcp") pod "1a204473-1a29-4367-af55-8acb6cf4aa17" (UID: "1a204473-1a29-4367-af55-8acb6cf4aa17"). InnerVolumeSpecName "kube-api-access-l4zcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:30:18 crc kubenswrapper[4926]: I1007 21:30:18.970464 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a204473-1a29-4367-af55-8acb6cf4aa17" (UID: "1a204473-1a29-4367-af55-8acb6cf4aa17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.030052 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.030081 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4zcp\" (UniqueName: \"kubernetes.io/projected/1a204473-1a29-4367-af55-8acb6cf4aa17-kube-api-access-l4zcp\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.030093 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a204473-1a29-4367-af55-8acb6cf4aa17-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.307226 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerID="5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51" exitCode=0 Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.307273 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerDied","Data":"5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51"} Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.307305 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlzfr" event={"ID":"1a204473-1a29-4367-af55-8acb6cf4aa17","Type":"ContainerDied","Data":"40782470f4499c2cc52a751d5de57b547c8883db9174c5612ebc6597d7fb5814"} Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.307325 4926 scope.go:117] "RemoveContainer" containerID="5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.307333 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlzfr" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.331428 4926 scope.go:117] "RemoveContainer" containerID="892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.360892 4926 scope.go:117] "RemoveContainer" containerID="cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.364103 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hlzfr"] Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.372663 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hlzfr"] Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.417384 4926 scope.go:117] "RemoveContainer" containerID="5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51" Oct 07 21:30:19 crc kubenswrapper[4926]: E1007 21:30:19.417808 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51\": container with ID starting with 5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51 not found: ID does not exist" containerID="5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.417863 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51"} err="failed to get container status \"5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51\": rpc error: code = NotFound desc = could not find container \"5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51\": container with ID starting with 5e70e6e234b7301d901651bac4160f19e2c596a4b560a7c7cc684f5c6df83e51 not found: ID does not exist" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.417904 4926 scope.go:117] "RemoveContainer" containerID="892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1" Oct 07 21:30:19 crc kubenswrapper[4926]: E1007 21:30:19.418583 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1\": container with ID starting with 892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1 not found: ID does not exist" containerID="892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.418654 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1"} err="failed to get container status \"892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1\": rpc error: code = NotFound desc = could not find container \"892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1\": container with ID starting with 892e265279bff3bd4d1133884cfa794a3956243eca865dc3f3428111ea7931f1 not found: ID does not exist" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.418694 4926 scope.go:117] "RemoveContainer" containerID="cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6" Oct 07 21:30:19 crc kubenswrapper[4926]: E1007 21:30:19.421542 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6\": container with ID starting with cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6 not found: ID does not exist" containerID="cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6" Oct 07 21:30:19 crc kubenswrapper[4926]: I1007 21:30:19.421616 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6"} err="failed to get container status \"cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6\": rpc error: code = NotFound desc = could not find container \"cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6\": container with ID starting with cf21995d75d908b6753d5b9bc20507480cd109c9d8fc89ed422aa9d58877ecc6 not found: ID does not exist" Oct 07 21:30:20 crc kubenswrapper[4926]: I1007 21:30:20.708314 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" path="/var/lib/kubelet/pods/1a204473-1a29-4367-af55-8acb6cf4aa17/volumes" Oct 07 21:31:22 crc kubenswrapper[4926]: I1007 21:31:22.032093 4926 generic.go:334] "Generic (PLEG): container finished" podID="bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" containerID="78ee9c2597196361ef2eecfa28478f0c4e4d15f045b13f3d22f17d62ff77c42d" exitCode=0 Oct 07 21:31:22 crc kubenswrapper[4926]: I1007 21:31:22.032172 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" event={"ID":"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12","Type":"ContainerDied","Data":"78ee9c2597196361ef2eecfa28478f0c4e4d15f045b13f3d22f17d62ff77c42d"} Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.518745 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.668705 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-inventory\") pod \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.668971 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdtc8\" (UniqueName: \"kubernetes.io/projected/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-kube-api-access-fdtc8\") pod \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.669227 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovncontroller-config-0\") pod \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.669285 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovn-combined-ca-bundle\") pod \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.669363 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ssh-key\") pod \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\" (UID: \"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12\") " Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.674161 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" (UID: "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.674522 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-kube-api-access-fdtc8" (OuterVolumeSpecName: "kube-api-access-fdtc8") pod "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" (UID: "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12"). InnerVolumeSpecName "kube-api-access-fdtc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.702867 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" (UID: "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.708610 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-inventory" (OuterVolumeSpecName: "inventory") pod "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" (UID: "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.723386 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" (UID: "bcecc1ad-1ce0-4b1a-8861-fc7d37276b12"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.771999 4926 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.772033 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.772042 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.772054 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:31:23 crc kubenswrapper[4926]: I1007 21:31:23.772062 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdtc8\" (UniqueName: \"kubernetes.io/projected/bcecc1ad-1ce0-4b1a-8861-fc7d37276b12-kube-api-access-fdtc8\") on node \"crc\" DevicePath \"\"" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.056682 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" event={"ID":"bcecc1ad-1ce0-4b1a-8861-fc7d37276b12","Type":"ContainerDied","Data":"9473bf20c5e79d9370068903ba58784402ed923d92868dcf6425b6af17d6a7ad"} Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.057099 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9473bf20c5e79d9370068903ba58784402ed923d92868dcf6425b6af17d6a7ad" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.057180 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwx7z" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.232573 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk"] Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233083 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233101 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233119 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="registry-server" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233127 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="registry-server" Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233143 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="extract-utilities" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233152 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="extract-utilities" Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233179 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="registry-server" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233187 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="registry-server" Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233224 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="extract-utilities" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233232 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="extract-utilities" Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233253 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="extract-content" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233261 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="extract-content" Oct 07 21:31:24 crc kubenswrapper[4926]: E1007 21:31:24.233275 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="extract-content" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233283 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="extract-content" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233523 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcecc1ad-1ce0-4b1a-8861-fc7d37276b12" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233550 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4d34cb-9dea-4f1f-9949-de74aa1d2ab2" containerName="registry-server" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.233570 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a204473-1a29-4367-af55-8acb6cf4aa17" containerName="registry-server" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.234396 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.236909 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.237217 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.237668 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.237931 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.238114 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.241768 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.257229 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk"] Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.391534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.391582 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.391612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.391780 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.391850 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdzzn\" (UniqueName: \"kubernetes.io/projected/bd775971-aba6-499a-b63f-ee121ded14d4-kube-api-access-bdzzn\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.392027 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.494045 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.494426 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.494568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.494767 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.494893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdzzn\" (UniqueName: \"kubernetes.io/projected/bd775971-aba6-499a-b63f-ee121ded14d4-kube-api-access-bdzzn\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.497034 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.499417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.499615 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.500663 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.505530 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.505900 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.525919 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdzzn\" (UniqueName: \"kubernetes.io/projected/bd775971-aba6-499a-b63f-ee121ded14d4-kube-api-access-bdzzn\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:24 crc kubenswrapper[4926]: I1007 21:31:24.553486 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:31:25 crc kubenswrapper[4926]: I1007 21:31:25.194161 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk"] Oct 07 21:31:25 crc kubenswrapper[4926]: W1007 21:31:25.196091 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd775971_aba6_499a_b63f_ee121ded14d4.slice/crio-d3fe1925bf64198828c3b29e19c6e75616da5d10c6b3e786a4c391e567eb36b5 WatchSource:0}: Error finding container d3fe1925bf64198828c3b29e19c6e75616da5d10c6b3e786a4c391e567eb36b5: Status 404 returned error can't find the container with id d3fe1925bf64198828c3b29e19c6e75616da5d10c6b3e786a4c391e567eb36b5 Oct 07 21:31:25 crc kubenswrapper[4926]: I1007 21:31:25.199494 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:31:26 crc kubenswrapper[4926]: I1007 21:31:26.080299 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" event={"ID":"bd775971-aba6-499a-b63f-ee121ded14d4","Type":"ContainerStarted","Data":"53ef2787114bc40f736047f9c54a85dc407386d407377f37a71749dc301054b1"} Oct 07 21:31:26 crc kubenswrapper[4926]: I1007 21:31:26.080849 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" event={"ID":"bd775971-aba6-499a-b63f-ee121ded14d4","Type":"ContainerStarted","Data":"d3fe1925bf64198828c3b29e19c6e75616da5d10c6b3e786a4c391e567eb36b5"} Oct 07 21:31:26 crc kubenswrapper[4926]: I1007 21:31:26.108554 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" podStartSLOduration=1.642820712 podStartE2EDuration="2.108529291s" podCreationTimestamp="2025-10-07 21:31:24 +0000 UTC" firstStartedPulling="2025-10-07 21:31:25.199277367 +0000 UTC m=+2135.237858517" lastFinishedPulling="2025-10-07 21:31:25.664985926 +0000 UTC m=+2135.703567096" observedRunningTime="2025-10-07 21:31:26.100639142 +0000 UTC m=+2136.139220292" watchObservedRunningTime="2025-10-07 21:31:26.108529291 +0000 UTC m=+2136.147110471" Oct 07 21:32:03 crc kubenswrapper[4926]: I1007 21:32:03.210060 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:32:03 crc kubenswrapper[4926]: I1007 21:32:03.211426 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:32:24 crc kubenswrapper[4926]: I1007 21:32:24.775279 4926 generic.go:334] "Generic (PLEG): container finished" podID="bd775971-aba6-499a-b63f-ee121ded14d4" containerID="53ef2787114bc40f736047f9c54a85dc407386d407377f37a71749dc301054b1" exitCode=0 Oct 07 21:32:24 crc kubenswrapper[4926]: I1007 21:32:24.775397 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" event={"ID":"bd775971-aba6-499a-b63f-ee121ded14d4","Type":"ContainerDied","Data":"53ef2787114bc40f736047f9c54a85dc407386d407377f37a71749dc301054b1"} Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.223796 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.350342 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-nova-metadata-neutron-config-0\") pod \"bd775971-aba6-499a-b63f-ee121ded14d4\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.350462 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-metadata-combined-ca-bundle\") pod \"bd775971-aba6-499a-b63f-ee121ded14d4\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.350554 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdzzn\" (UniqueName: \"kubernetes.io/projected/bd775971-aba6-499a-b63f-ee121ded14d4-kube-api-access-bdzzn\") pod \"bd775971-aba6-499a-b63f-ee121ded14d4\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.350586 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-ssh-key\") pod \"bd775971-aba6-499a-b63f-ee121ded14d4\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.350660 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"bd775971-aba6-499a-b63f-ee121ded14d4\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.350707 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-inventory\") pod \"bd775971-aba6-499a-b63f-ee121ded14d4\" (UID: \"bd775971-aba6-499a-b63f-ee121ded14d4\") " Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.356402 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd775971-aba6-499a-b63f-ee121ded14d4-kube-api-access-bdzzn" (OuterVolumeSpecName: "kube-api-access-bdzzn") pod "bd775971-aba6-499a-b63f-ee121ded14d4" (UID: "bd775971-aba6-499a-b63f-ee121ded14d4"). InnerVolumeSpecName "kube-api-access-bdzzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.360619 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "bd775971-aba6-499a-b63f-ee121ded14d4" (UID: "bd775971-aba6-499a-b63f-ee121ded14d4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.378624 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-inventory" (OuterVolumeSpecName: "inventory") pod "bd775971-aba6-499a-b63f-ee121ded14d4" (UID: "bd775971-aba6-499a-b63f-ee121ded14d4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.380754 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "bd775971-aba6-499a-b63f-ee121ded14d4" (UID: "bd775971-aba6-499a-b63f-ee121ded14d4"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.398106 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd775971-aba6-499a-b63f-ee121ded14d4" (UID: "bd775971-aba6-499a-b63f-ee121ded14d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.416631 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "bd775971-aba6-499a-b63f-ee121ded14d4" (UID: "bd775971-aba6-499a-b63f-ee121ded14d4"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.454133 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdzzn\" (UniqueName: \"kubernetes.io/projected/bd775971-aba6-499a-b63f-ee121ded14d4-kube-api-access-bdzzn\") on node \"crc\" DevicePath \"\"" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.454172 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.454188 4926 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.454221 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.454233 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.454249 4926 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd775971-aba6-499a-b63f-ee121ded14d4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.799386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" event={"ID":"bd775971-aba6-499a-b63f-ee121ded14d4","Type":"ContainerDied","Data":"d3fe1925bf64198828c3b29e19c6e75616da5d10c6b3e786a4c391e567eb36b5"} Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.799458 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3fe1925bf64198828c3b29e19c6e75616da5d10c6b3e786a4c391e567eb36b5" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.799483 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.934624 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr"] Oct 07 21:32:26 crc kubenswrapper[4926]: E1007 21:32:26.935371 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd775971-aba6-499a-b63f-ee121ded14d4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.935406 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd775971-aba6-499a-b63f-ee121ded14d4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.935749 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd775971-aba6-499a-b63f-ee121ded14d4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.936950 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.940677 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.941029 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.941106 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.941341 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.941618 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.948309 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr"] Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.966165 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.966302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwf7b\" (UniqueName: \"kubernetes.io/projected/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-kube-api-access-dwf7b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.966377 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.966435 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:26 crc kubenswrapper[4926]: I1007 21:32:26.966647 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.068549 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.068699 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.068758 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.068821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwf7b\" (UniqueName: \"kubernetes.io/projected/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-kube-api-access-dwf7b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.068864 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.077754 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.079108 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.082911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.088022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.091751 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwf7b\" (UniqueName: \"kubernetes.io/projected/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-kube-api-access-dwf7b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.262904 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:32:27 crc kubenswrapper[4926]: I1007 21:32:27.824322 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr"] Oct 07 21:32:28 crc kubenswrapper[4926]: I1007 21:32:28.823662 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" event={"ID":"4c8adb00-aa62-4d33-9b2f-a9b19b57e893","Type":"ContainerStarted","Data":"99aca347ae453039430e2d535be940390397d962e6aa5ff59a0fa93597b75405"} Oct 07 21:32:28 crc kubenswrapper[4926]: I1007 21:32:28.824092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" event={"ID":"4c8adb00-aa62-4d33-9b2f-a9b19b57e893","Type":"ContainerStarted","Data":"edeac45c2462631c8dcc6aa14b5ac08d4c9d7faee0edae403bfad67cf09f8a43"} Oct 07 21:32:28 crc kubenswrapper[4926]: I1007 21:32:28.851298 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" podStartSLOduration=2.349127841 podStartE2EDuration="2.851276338s" podCreationTimestamp="2025-10-07 21:32:26 +0000 UTC" firstStartedPulling="2025-10-07 21:32:27.835798828 +0000 UTC m=+2197.874379978" lastFinishedPulling="2025-10-07 21:32:28.337947325 +0000 UTC m=+2198.376528475" observedRunningTime="2025-10-07 21:32:28.848886768 +0000 UTC m=+2198.887467928" watchObservedRunningTime="2025-10-07 21:32:28.851276338 +0000 UTC m=+2198.889857488" Oct 07 21:32:33 crc kubenswrapper[4926]: I1007 21:32:33.210403 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:32:33 crc kubenswrapper[4926]: I1007 21:32:33.211148 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:33:03 crc kubenswrapper[4926]: I1007 21:33:03.209922 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:33:03 crc kubenswrapper[4926]: I1007 21:33:03.210559 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:33:03 crc kubenswrapper[4926]: I1007 21:33:03.210611 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:33:03 crc kubenswrapper[4926]: I1007 21:33:03.211599 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:33:03 crc kubenswrapper[4926]: I1007 21:33:03.211701 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" gracePeriod=600 Oct 07 21:33:03 crc kubenswrapper[4926]: E1007 21:33:03.346569 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:33:04 crc kubenswrapper[4926]: I1007 21:33:04.279172 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" exitCode=0 Oct 07 21:33:04 crc kubenswrapper[4926]: I1007 21:33:04.279263 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8"} Oct 07 21:33:04 crc kubenswrapper[4926]: I1007 21:33:04.279602 4926 scope.go:117] "RemoveContainer" containerID="b0fadebd37b106a55391d1cc2cbe36eb8d7a6c92c1c552b732e0960ab706c9f6" Oct 07 21:33:04 crc kubenswrapper[4926]: I1007 21:33:04.280881 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:33:04 crc kubenswrapper[4926]: E1007 21:33:04.281556 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:33:17 crc kubenswrapper[4926]: I1007 21:33:17.679268 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:33:17 crc kubenswrapper[4926]: E1007 21:33:17.679970 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:33:28 crc kubenswrapper[4926]: I1007 21:33:28.681407 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:33:28 crc kubenswrapper[4926]: E1007 21:33:28.682270 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:33:42 crc kubenswrapper[4926]: I1007 21:33:42.679330 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:33:42 crc kubenswrapper[4926]: E1007 21:33:42.680616 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:33:57 crc kubenswrapper[4926]: I1007 21:33:57.679631 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:33:57 crc kubenswrapper[4926]: E1007 21:33:57.680647 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:34:11 crc kubenswrapper[4926]: I1007 21:34:11.679009 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:34:11 crc kubenswrapper[4926]: E1007 21:34:11.680109 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:34:23 crc kubenswrapper[4926]: I1007 21:34:23.679108 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:34:23 crc kubenswrapper[4926]: E1007 21:34:23.680089 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.318914 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rc22m"] Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.323144 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.333875 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rc22m"] Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.434502 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-utilities\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.434573 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgkh4\" (UniqueName: \"kubernetes.io/projected/513403f7-f627-4b42-9890-c55c879e1007-kube-api-access-dgkh4\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.434730 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-catalog-content\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.537614 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-utilities\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.537680 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgkh4\" (UniqueName: \"kubernetes.io/projected/513403f7-f627-4b42-9890-c55c879e1007-kube-api-access-dgkh4\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.537749 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-catalog-content\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.538266 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-catalog-content\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.538485 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-utilities\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.559756 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgkh4\" (UniqueName: \"kubernetes.io/projected/513403f7-f627-4b42-9890-c55c879e1007-kube-api-access-dgkh4\") pod \"community-operators-rc22m\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:27 crc kubenswrapper[4926]: I1007 21:34:27.653491 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:28 crc kubenswrapper[4926]: I1007 21:34:28.209540 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rc22m"] Oct 07 21:34:28 crc kubenswrapper[4926]: I1007 21:34:28.278428 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc22m" event={"ID":"513403f7-f627-4b42-9890-c55c879e1007","Type":"ContainerStarted","Data":"b64fe29d320af56c3e6534972dbf09cb988897a56c93b4b0ae2375a027a115d4"} Oct 07 21:34:29 crc kubenswrapper[4926]: I1007 21:34:29.288685 4926 generic.go:334] "Generic (PLEG): container finished" podID="513403f7-f627-4b42-9890-c55c879e1007" containerID="cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852" exitCode=0 Oct 07 21:34:29 crc kubenswrapper[4926]: I1007 21:34:29.288808 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc22m" event={"ID":"513403f7-f627-4b42-9890-c55c879e1007","Type":"ContainerDied","Data":"cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852"} Oct 07 21:34:30 crc kubenswrapper[4926]: I1007 21:34:30.298230 4926 generic.go:334] "Generic (PLEG): container finished" podID="513403f7-f627-4b42-9890-c55c879e1007" containerID="f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9" exitCode=0 Oct 07 21:34:30 crc kubenswrapper[4926]: I1007 21:34:30.298308 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc22m" event={"ID":"513403f7-f627-4b42-9890-c55c879e1007","Type":"ContainerDied","Data":"f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9"} Oct 07 21:34:31 crc kubenswrapper[4926]: I1007 21:34:31.309500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc22m" event={"ID":"513403f7-f627-4b42-9890-c55c879e1007","Type":"ContainerStarted","Data":"090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186"} Oct 07 21:34:31 crc kubenswrapper[4926]: I1007 21:34:31.336377 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rc22m" podStartSLOduration=2.7472595220000002 podStartE2EDuration="4.336351943s" podCreationTimestamp="2025-10-07 21:34:27 +0000 UTC" firstStartedPulling="2025-10-07 21:34:29.291752134 +0000 UTC m=+2319.330333284" lastFinishedPulling="2025-10-07 21:34:30.880844545 +0000 UTC m=+2320.919425705" observedRunningTime="2025-10-07 21:34:31.334288313 +0000 UTC m=+2321.372869473" watchObservedRunningTime="2025-10-07 21:34:31.336351943 +0000 UTC m=+2321.374933113" Oct 07 21:34:34 crc kubenswrapper[4926]: I1007 21:34:34.679565 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:34:34 crc kubenswrapper[4926]: E1007 21:34:34.680133 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:34:37 crc kubenswrapper[4926]: I1007 21:34:37.654611 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:37 crc kubenswrapper[4926]: I1007 21:34:37.655167 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:37 crc kubenswrapper[4926]: I1007 21:34:37.732405 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:38 crc kubenswrapper[4926]: I1007 21:34:38.423521 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:38 crc kubenswrapper[4926]: I1007 21:34:38.476810 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rc22m"] Oct 07 21:34:40 crc kubenswrapper[4926]: I1007 21:34:40.416784 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rc22m" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="registry-server" containerID="cri-o://090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186" gracePeriod=2 Oct 07 21:34:40 crc kubenswrapper[4926]: I1007 21:34:40.956294 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.153592 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-utilities\") pod \"513403f7-f627-4b42-9890-c55c879e1007\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.153839 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgkh4\" (UniqueName: \"kubernetes.io/projected/513403f7-f627-4b42-9890-c55c879e1007-kube-api-access-dgkh4\") pod \"513403f7-f627-4b42-9890-c55c879e1007\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.153948 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-catalog-content\") pod \"513403f7-f627-4b42-9890-c55c879e1007\" (UID: \"513403f7-f627-4b42-9890-c55c879e1007\") " Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.154463 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-utilities" (OuterVolumeSpecName: "utilities") pod "513403f7-f627-4b42-9890-c55c879e1007" (UID: "513403f7-f627-4b42-9890-c55c879e1007"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.155241 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.160852 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/513403f7-f627-4b42-9890-c55c879e1007-kube-api-access-dgkh4" (OuterVolumeSpecName: "kube-api-access-dgkh4") pod "513403f7-f627-4b42-9890-c55c879e1007" (UID: "513403f7-f627-4b42-9890-c55c879e1007"). InnerVolumeSpecName "kube-api-access-dgkh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.214923 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "513403f7-f627-4b42-9890-c55c879e1007" (UID: "513403f7-f627-4b42-9890-c55c879e1007"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.257281 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgkh4\" (UniqueName: \"kubernetes.io/projected/513403f7-f627-4b42-9890-c55c879e1007-kube-api-access-dgkh4\") on node \"crc\" DevicePath \"\"" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.257309 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/513403f7-f627-4b42-9890-c55c879e1007-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.432887 4926 generic.go:334] "Generic (PLEG): container finished" podID="513403f7-f627-4b42-9890-c55c879e1007" containerID="090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186" exitCode=0 Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.432949 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc22m" event={"ID":"513403f7-f627-4b42-9890-c55c879e1007","Type":"ContainerDied","Data":"090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186"} Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.433002 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc22m" event={"ID":"513403f7-f627-4b42-9890-c55c879e1007","Type":"ContainerDied","Data":"b64fe29d320af56c3e6534972dbf09cb988897a56c93b4b0ae2375a027a115d4"} Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.433034 4926 scope.go:117] "RemoveContainer" containerID="090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.433289 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc22m" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.464405 4926 scope.go:117] "RemoveContainer" containerID="f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.498749 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rc22m"] Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.499009 4926 scope.go:117] "RemoveContainer" containerID="cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.513242 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rc22m"] Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.565173 4926 scope.go:117] "RemoveContainer" containerID="090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186" Oct 07 21:34:41 crc kubenswrapper[4926]: E1007 21:34:41.565923 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186\": container with ID starting with 090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186 not found: ID does not exist" containerID="090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.565964 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186"} err="failed to get container status \"090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186\": rpc error: code = NotFound desc = could not find container \"090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186\": container with ID starting with 090b1398df1707c23a43e6c10a4dfd1945c91e15296823762e42592f1bdad186 not found: ID does not exist" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.566015 4926 scope.go:117] "RemoveContainer" containerID="f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9" Oct 07 21:34:41 crc kubenswrapper[4926]: E1007 21:34:41.566506 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9\": container with ID starting with f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9 not found: ID does not exist" containerID="f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.566538 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9"} err="failed to get container status \"f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9\": rpc error: code = NotFound desc = could not find container \"f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9\": container with ID starting with f6d27338d7c122bac68913f8e3a27f3def3056e71dd462392f61c676593de8f9 not found: ID does not exist" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.566562 4926 scope.go:117] "RemoveContainer" containerID="cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852" Oct 07 21:34:41 crc kubenswrapper[4926]: E1007 21:34:41.566854 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852\": container with ID starting with cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852 not found: ID does not exist" containerID="cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852" Oct 07 21:34:41 crc kubenswrapper[4926]: I1007 21:34:41.566885 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852"} err="failed to get container status \"cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852\": rpc error: code = NotFound desc = could not find container \"cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852\": container with ID starting with cf774c39df4049a2df364fa87efd8e28deee3ea1a5db6c68d37284b6268a3852 not found: ID does not exist" Oct 07 21:34:42 crc kubenswrapper[4926]: I1007 21:34:42.693849 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="513403f7-f627-4b42-9890-c55c879e1007" path="/var/lib/kubelet/pods/513403f7-f627-4b42-9890-c55c879e1007/volumes" Oct 07 21:34:48 crc kubenswrapper[4926]: I1007 21:34:48.680157 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:34:48 crc kubenswrapper[4926]: E1007 21:34:48.681047 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:35:00 crc kubenswrapper[4926]: I1007 21:35:00.678680 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:35:00 crc kubenswrapper[4926]: E1007 21:35:00.679634 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:35:11 crc kubenswrapper[4926]: I1007 21:35:11.680469 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:35:11 crc kubenswrapper[4926]: E1007 21:35:11.682741 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.679943 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:35:22 crc kubenswrapper[4926]: E1007 21:35:22.681117 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.969786 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gxstq"] Oct 07 21:35:22 crc kubenswrapper[4926]: E1007 21:35:22.970525 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="registry-server" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.970556 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="registry-server" Oct 07 21:35:22 crc kubenswrapper[4926]: E1007 21:35:22.970622 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="extract-content" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.970635 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="extract-content" Oct 07 21:35:22 crc kubenswrapper[4926]: E1007 21:35:22.970657 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="extract-utilities" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.970670 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="extract-utilities" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.971005 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="513403f7-f627-4b42-9890-c55c879e1007" containerName="registry-server" Oct 07 21:35:22 crc kubenswrapper[4926]: I1007 21:35:22.973615 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.001620 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxstq"] Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.096403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4thrm\" (UniqueName: \"kubernetes.io/projected/d85c2244-8d30-4350-8770-e640b8a862bf-kube-api-access-4thrm\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.096735 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-catalog-content\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.096779 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-utilities\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.198807 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4thrm\" (UniqueName: \"kubernetes.io/projected/d85c2244-8d30-4350-8770-e640b8a862bf-kube-api-access-4thrm\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.198858 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-catalog-content\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.198912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-utilities\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.199492 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-utilities\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.199946 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-catalog-content\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.222707 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4thrm\" (UniqueName: \"kubernetes.io/projected/d85c2244-8d30-4350-8770-e640b8a862bf-kube-api-access-4thrm\") pod \"redhat-marketplace-gxstq\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.320957 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.842602 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxstq"] Oct 07 21:35:23 crc kubenswrapper[4926]: I1007 21:35:23.948101 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxstq" event={"ID":"d85c2244-8d30-4350-8770-e640b8a862bf","Type":"ContainerStarted","Data":"da250ba8efbd4c907f618750eb088cbc54a9f166334d13a2857f9e77cdb813d2"} Oct 07 21:35:24 crc kubenswrapper[4926]: I1007 21:35:24.962262 4926 generic.go:334] "Generic (PLEG): container finished" podID="d85c2244-8d30-4350-8770-e640b8a862bf" containerID="db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4" exitCode=0 Oct 07 21:35:24 crc kubenswrapper[4926]: I1007 21:35:24.962374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxstq" event={"ID":"d85c2244-8d30-4350-8770-e640b8a862bf","Type":"ContainerDied","Data":"db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4"} Oct 07 21:35:25 crc kubenswrapper[4926]: I1007 21:35:25.974001 4926 generic.go:334] "Generic (PLEG): container finished" podID="d85c2244-8d30-4350-8770-e640b8a862bf" containerID="616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede" exitCode=0 Oct 07 21:35:25 crc kubenswrapper[4926]: I1007 21:35:25.974375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxstq" event={"ID":"d85c2244-8d30-4350-8770-e640b8a862bf","Type":"ContainerDied","Data":"616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede"} Oct 07 21:35:26 crc kubenswrapper[4926]: I1007 21:35:26.986813 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxstq" event={"ID":"d85c2244-8d30-4350-8770-e640b8a862bf","Type":"ContainerStarted","Data":"99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b"} Oct 07 21:35:27 crc kubenswrapper[4926]: I1007 21:35:27.005282 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gxstq" podStartSLOduration=3.528973957 podStartE2EDuration="5.005260251s" podCreationTimestamp="2025-10-07 21:35:22 +0000 UTC" firstStartedPulling="2025-10-07 21:35:24.967946225 +0000 UTC m=+2375.006527425" lastFinishedPulling="2025-10-07 21:35:26.444232559 +0000 UTC m=+2376.482813719" observedRunningTime="2025-10-07 21:35:27.004902991 +0000 UTC m=+2377.043484141" watchObservedRunningTime="2025-10-07 21:35:27.005260251 +0000 UTC m=+2377.043841411" Oct 07 21:35:33 crc kubenswrapper[4926]: I1007 21:35:33.321337 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:33 crc kubenswrapper[4926]: I1007 21:35:33.322123 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:33 crc kubenswrapper[4926]: I1007 21:35:33.395981 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:34 crc kubenswrapper[4926]: I1007 21:35:34.131385 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:34 crc kubenswrapper[4926]: I1007 21:35:34.203853 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxstq"] Oct 07 21:35:34 crc kubenswrapper[4926]: I1007 21:35:34.680903 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:35:34 crc kubenswrapper[4926]: E1007 21:35:34.681345 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.097498 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gxstq" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="registry-server" containerID="cri-o://99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b" gracePeriod=2 Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.598548 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.725092 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4thrm\" (UniqueName: \"kubernetes.io/projected/d85c2244-8d30-4350-8770-e640b8a862bf-kube-api-access-4thrm\") pod \"d85c2244-8d30-4350-8770-e640b8a862bf\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.725896 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-catalog-content\") pod \"d85c2244-8d30-4350-8770-e640b8a862bf\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.726025 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-utilities\") pod \"d85c2244-8d30-4350-8770-e640b8a862bf\" (UID: \"d85c2244-8d30-4350-8770-e640b8a862bf\") " Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.727095 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-utilities" (OuterVolumeSpecName: "utilities") pod "d85c2244-8d30-4350-8770-e640b8a862bf" (UID: "d85c2244-8d30-4350-8770-e640b8a862bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.734867 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d85c2244-8d30-4350-8770-e640b8a862bf-kube-api-access-4thrm" (OuterVolumeSpecName: "kube-api-access-4thrm") pod "d85c2244-8d30-4350-8770-e640b8a862bf" (UID: "d85c2244-8d30-4350-8770-e640b8a862bf"). InnerVolumeSpecName "kube-api-access-4thrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.741527 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d85c2244-8d30-4350-8770-e640b8a862bf" (UID: "d85c2244-8d30-4350-8770-e640b8a862bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.828905 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.829337 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4thrm\" (UniqueName: \"kubernetes.io/projected/d85c2244-8d30-4350-8770-e640b8a862bf-kube-api-access-4thrm\") on node \"crc\" DevicePath \"\"" Oct 07 21:35:36 crc kubenswrapper[4926]: I1007 21:35:36.829505 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d85c2244-8d30-4350-8770-e640b8a862bf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.112422 4926 generic.go:334] "Generic (PLEG): container finished" podID="d85c2244-8d30-4350-8770-e640b8a862bf" containerID="99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b" exitCode=0 Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.112478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxstq" event={"ID":"d85c2244-8d30-4350-8770-e640b8a862bf","Type":"ContainerDied","Data":"99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b"} Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.112527 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxstq" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.112554 4926 scope.go:117] "RemoveContainer" containerID="99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.112539 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxstq" event={"ID":"d85c2244-8d30-4350-8770-e640b8a862bf","Type":"ContainerDied","Data":"da250ba8efbd4c907f618750eb088cbc54a9f166334d13a2857f9e77cdb813d2"} Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.164777 4926 scope.go:117] "RemoveContainer" containerID="616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.179579 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxstq"] Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.189542 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxstq"] Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.198724 4926 scope.go:117] "RemoveContainer" containerID="db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.258083 4926 scope.go:117] "RemoveContainer" containerID="99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b" Oct 07 21:35:37 crc kubenswrapper[4926]: E1007 21:35:37.258572 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b\": container with ID starting with 99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b not found: ID does not exist" containerID="99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.258621 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b"} err="failed to get container status \"99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b\": rpc error: code = NotFound desc = could not find container \"99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b\": container with ID starting with 99921b85d05b710f270d3a42041e84e1539b5a79df6bf9dc167b5633ffe3ba5b not found: ID does not exist" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.258657 4926 scope.go:117] "RemoveContainer" containerID="616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede" Oct 07 21:35:37 crc kubenswrapper[4926]: E1007 21:35:37.259154 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede\": container with ID starting with 616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede not found: ID does not exist" containerID="616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.259227 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede"} err="failed to get container status \"616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede\": rpc error: code = NotFound desc = could not find container \"616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede\": container with ID starting with 616fb25f28e1a7a73fc357d8ec1b877108a25f2cf841aaf4eba2ceb575318ede not found: ID does not exist" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.259266 4926 scope.go:117] "RemoveContainer" containerID="db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4" Oct 07 21:35:37 crc kubenswrapper[4926]: E1007 21:35:37.259834 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4\": container with ID starting with db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4 not found: ID does not exist" containerID="db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4" Oct 07 21:35:37 crc kubenswrapper[4926]: I1007 21:35:37.259878 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4"} err="failed to get container status \"db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4\": rpc error: code = NotFound desc = could not find container \"db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4\": container with ID starting with db385e5b592455d535459c024c7f95f796082902e5c3618cdbf54c1d7f4bf1c4 not found: ID does not exist" Oct 07 21:35:38 crc kubenswrapper[4926]: I1007 21:35:38.696299 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" path="/var/lib/kubelet/pods/d85c2244-8d30-4350-8770-e640b8a862bf/volumes" Oct 07 21:35:49 crc kubenswrapper[4926]: I1007 21:35:49.679565 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:35:49 crc kubenswrapper[4926]: E1007 21:35:49.680728 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:36:04 crc kubenswrapper[4926]: I1007 21:36:04.679828 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:36:04 crc kubenswrapper[4926]: E1007 21:36:04.681115 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:36:17 crc kubenswrapper[4926]: I1007 21:36:17.679405 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:36:17 crc kubenswrapper[4926]: E1007 21:36:17.679995 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:36:31 crc kubenswrapper[4926]: I1007 21:36:31.679615 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:36:31 crc kubenswrapper[4926]: E1007 21:36:31.680506 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:36:42 crc kubenswrapper[4926]: I1007 21:36:42.679857 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:36:42 crc kubenswrapper[4926]: E1007 21:36:42.680822 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:36:54 crc kubenswrapper[4926]: I1007 21:36:54.680394 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:36:54 crc kubenswrapper[4926]: E1007 21:36:54.681726 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:37:05 crc kubenswrapper[4926]: I1007 21:37:05.679617 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:37:05 crc kubenswrapper[4926]: E1007 21:37:05.680380 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:37:17 crc kubenswrapper[4926]: I1007 21:37:17.198592 4926 generic.go:334] "Generic (PLEG): container finished" podID="4c8adb00-aa62-4d33-9b2f-a9b19b57e893" containerID="99aca347ae453039430e2d535be940390397d962e6aa5ff59a0fa93597b75405" exitCode=0 Oct 07 21:37:17 crc kubenswrapper[4926]: I1007 21:37:17.198723 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" event={"ID":"4c8adb00-aa62-4d33-9b2f-a9b19b57e893","Type":"ContainerDied","Data":"99aca347ae453039430e2d535be940390397d962e6aa5ff59a0fa93597b75405"} Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.754506 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.879136 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-inventory\") pod \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.879224 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-combined-ca-bundle\") pod \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.879296 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-ssh-key\") pod \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.879421 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-secret-0\") pod \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.879482 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwf7b\" (UniqueName: \"kubernetes.io/projected/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-kube-api-access-dwf7b\") pod \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\" (UID: \"4c8adb00-aa62-4d33-9b2f-a9b19b57e893\") " Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.886732 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-kube-api-access-dwf7b" (OuterVolumeSpecName: "kube-api-access-dwf7b") pod "4c8adb00-aa62-4d33-9b2f-a9b19b57e893" (UID: "4c8adb00-aa62-4d33-9b2f-a9b19b57e893"). InnerVolumeSpecName "kube-api-access-dwf7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.886884 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4c8adb00-aa62-4d33-9b2f-a9b19b57e893" (UID: "4c8adb00-aa62-4d33-9b2f-a9b19b57e893"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.911860 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4c8adb00-aa62-4d33-9b2f-a9b19b57e893" (UID: "4c8adb00-aa62-4d33-9b2f-a9b19b57e893"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.911937 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4c8adb00-aa62-4d33-9b2f-a9b19b57e893" (UID: "4c8adb00-aa62-4d33-9b2f-a9b19b57e893"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.929678 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-inventory" (OuterVolumeSpecName: "inventory") pod "4c8adb00-aa62-4d33-9b2f-a9b19b57e893" (UID: "4c8adb00-aa62-4d33-9b2f-a9b19b57e893"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.982010 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.982041 4926 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.982052 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.982062 4926 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:37:18 crc kubenswrapper[4926]: I1007 21:37:18.982072 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwf7b\" (UniqueName: \"kubernetes.io/projected/4c8adb00-aa62-4d33-9b2f-a9b19b57e893-kube-api-access-dwf7b\") on node \"crc\" DevicePath \"\"" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.219818 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" event={"ID":"4c8adb00-aa62-4d33-9b2f-a9b19b57e893","Type":"ContainerDied","Data":"edeac45c2462631c8dcc6aa14b5ac08d4c9d7faee0edae403bfad67cf09f8a43"} Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.219862 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edeac45c2462631c8dcc6aa14b5ac08d4c9d7faee0edae403bfad67cf09f8a43" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.219924 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.331236 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb"] Oct 07 21:37:19 crc kubenswrapper[4926]: E1007 21:37:19.331628 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="registry-server" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.331659 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="registry-server" Oct 07 21:37:19 crc kubenswrapper[4926]: E1007 21:37:19.331684 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8adb00-aa62-4d33-9b2f-a9b19b57e893" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.331693 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8adb00-aa62-4d33-9b2f-a9b19b57e893" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 07 21:37:19 crc kubenswrapper[4926]: E1007 21:37:19.331712 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="extract-utilities" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.331719 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="extract-utilities" Oct 07 21:37:19 crc kubenswrapper[4926]: E1007 21:37:19.331740 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="extract-content" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.331747 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="extract-content" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.331969 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d85c2244-8d30-4350-8770-e640b8a862bf" containerName="registry-server" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.332014 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8adb00-aa62-4d33-9b2f-a9b19b57e893" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.333062 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.338945 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.339080 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.342721 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.342918 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.343002 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.343049 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.343101 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.368170 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb"] Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390211 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmczx\" (UniqueName: \"kubernetes.io/projected/0273b882-c15f-4ca8-b786-60fb42c205ab-kube-api-access-xmczx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390684 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390737 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390776 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390835 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.390895 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.391042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.391085 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492760 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492824 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492860 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492914 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492974 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmczx\" (UniqueName: \"kubernetes.io/projected/0273b882-c15f-4ca8-b786-60fb42c205ab-kube-api-access-xmczx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.492998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.493050 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.493072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.494709 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.497083 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.497119 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.498774 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.498920 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.499353 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.499373 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.501774 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.510990 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmczx\" (UniqueName: \"kubernetes.io/projected/0273b882-c15f-4ca8-b786-60fb42c205ab-kube-api-access-xmczx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t7hkb\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:19 crc kubenswrapper[4926]: I1007 21:37:19.658985 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:37:20 crc kubenswrapper[4926]: I1007 21:37:20.216039 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb"] Oct 07 21:37:20 crc kubenswrapper[4926]: I1007 21:37:20.220816 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:37:20 crc kubenswrapper[4926]: I1007 21:37:20.233324 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" event={"ID":"0273b882-c15f-4ca8-b786-60fb42c205ab","Type":"ContainerStarted","Data":"be7ed188ead179f7ba7c656bf1f9863c3357003b7ec2e22385c09f56d6f4d35c"} Oct 07 21:37:20 crc kubenswrapper[4926]: I1007 21:37:20.695919 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:37:20 crc kubenswrapper[4926]: E1007 21:37:20.696577 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:37:21 crc kubenswrapper[4926]: I1007 21:37:21.242335 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" event={"ID":"0273b882-c15f-4ca8-b786-60fb42c205ab","Type":"ContainerStarted","Data":"fc12c5b185971822d0fa3570f5befe43c57940f0163b192eca577c66bbb59494"} Oct 07 21:37:35 crc kubenswrapper[4926]: I1007 21:37:35.679697 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:37:35 crc kubenswrapper[4926]: E1007 21:37:35.680374 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:37:48 crc kubenswrapper[4926]: I1007 21:37:48.679442 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:37:48 crc kubenswrapper[4926]: E1007 21:37:48.680386 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:38:02 crc kubenswrapper[4926]: I1007 21:38:02.679130 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:38:02 crc kubenswrapper[4926]: E1007 21:38:02.680170 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:38:13 crc kubenswrapper[4926]: I1007 21:38:13.678911 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:38:14 crc kubenswrapper[4926]: I1007 21:38:14.841946 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"45407472135a43dc033ae09daa12cedb15f24f7ec62fc31880605bb03ece9f39"} Oct 07 21:38:14 crc kubenswrapper[4926]: I1007 21:38:14.879040 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" podStartSLOduration=55.319493175 podStartE2EDuration="55.879005604s" podCreationTimestamp="2025-10-07 21:37:19 +0000 UTC" firstStartedPulling="2025-10-07 21:37:20.220450137 +0000 UTC m=+2490.259031297" lastFinishedPulling="2025-10-07 21:37:20.779962566 +0000 UTC m=+2490.818543726" observedRunningTime="2025-10-07 21:37:21.261586538 +0000 UTC m=+2491.300167688" watchObservedRunningTime="2025-10-07 21:38:14.879005604 +0000 UTC m=+2544.917586824" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.130927 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jpxpq"] Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.136919 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.171146 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jpxpq"] Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.221699 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-catalog-content\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.221782 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4t6s\" (UniqueName: \"kubernetes.io/projected/6d990095-cc8f-479e-95a6-0c90ad349aab-kube-api-access-r4t6s\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.221844 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-utilities\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.324482 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-catalog-content\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.324543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4t6s\" (UniqueName: \"kubernetes.io/projected/6d990095-cc8f-479e-95a6-0c90ad349aab-kube-api-access-r4t6s\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.324590 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-utilities\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.325054 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-catalog-content\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.325218 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-utilities\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.361477 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4t6s\" (UniqueName: \"kubernetes.io/projected/6d990095-cc8f-479e-95a6-0c90ad349aab-kube-api-access-r4t6s\") pod \"certified-operators-jpxpq\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:05 crc kubenswrapper[4926]: I1007 21:40:05.473524 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:06 crc kubenswrapper[4926]: I1007 21:40:06.072445 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jpxpq"] Oct 07 21:40:06 crc kubenswrapper[4926]: I1007 21:40:06.110447 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jpxpq" event={"ID":"6d990095-cc8f-479e-95a6-0c90ad349aab","Type":"ContainerStarted","Data":"80a14711614f13023709f1045b7322f91500613a53a9a2809b5f92bd7117ce18"} Oct 07 21:40:07 crc kubenswrapper[4926]: I1007 21:40:07.142306 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerID="138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933" exitCode=0 Oct 07 21:40:07 crc kubenswrapper[4926]: I1007 21:40:07.142499 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jpxpq" event={"ID":"6d990095-cc8f-479e-95a6-0c90ad349aab","Type":"ContainerDied","Data":"138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933"} Oct 07 21:40:09 crc kubenswrapper[4926]: I1007 21:40:09.166709 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerID="762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac" exitCode=0 Oct 07 21:40:09 crc kubenswrapper[4926]: I1007 21:40:09.166747 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jpxpq" event={"ID":"6d990095-cc8f-479e-95a6-0c90ad349aab","Type":"ContainerDied","Data":"762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac"} Oct 07 21:40:10 crc kubenswrapper[4926]: I1007 21:40:10.182403 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jpxpq" event={"ID":"6d990095-cc8f-479e-95a6-0c90ad349aab","Type":"ContainerStarted","Data":"db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a"} Oct 07 21:40:10 crc kubenswrapper[4926]: I1007 21:40:10.227624 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jpxpq" podStartSLOduration=2.544123721 podStartE2EDuration="5.22760287s" podCreationTimestamp="2025-10-07 21:40:05 +0000 UTC" firstStartedPulling="2025-10-07 21:40:07.145649581 +0000 UTC m=+2657.184230771" lastFinishedPulling="2025-10-07 21:40:09.82912876 +0000 UTC m=+2659.867709920" observedRunningTime="2025-10-07 21:40:10.214526841 +0000 UTC m=+2660.253108001" watchObservedRunningTime="2025-10-07 21:40:10.22760287 +0000 UTC m=+2660.266184030" Oct 07 21:40:15 crc kubenswrapper[4926]: I1007 21:40:15.475109 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:15 crc kubenswrapper[4926]: I1007 21:40:15.475615 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:15 crc kubenswrapper[4926]: I1007 21:40:15.533168 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:16 crc kubenswrapper[4926]: I1007 21:40:16.338555 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:16 crc kubenswrapper[4926]: I1007 21:40:16.398610 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jpxpq"] Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.283528 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jpxpq" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="registry-server" containerID="cri-o://db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a" gracePeriod=2 Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.744602 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.841574 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-utilities\") pod \"6d990095-cc8f-479e-95a6-0c90ad349aab\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.841807 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4t6s\" (UniqueName: \"kubernetes.io/projected/6d990095-cc8f-479e-95a6-0c90ad349aab-kube-api-access-r4t6s\") pod \"6d990095-cc8f-479e-95a6-0c90ad349aab\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.841851 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-catalog-content\") pod \"6d990095-cc8f-479e-95a6-0c90ad349aab\" (UID: \"6d990095-cc8f-479e-95a6-0c90ad349aab\") " Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.843104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-utilities" (OuterVolumeSpecName: "utilities") pod "6d990095-cc8f-479e-95a6-0c90ad349aab" (UID: "6d990095-cc8f-479e-95a6-0c90ad349aab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.848898 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d990095-cc8f-479e-95a6-0c90ad349aab-kube-api-access-r4t6s" (OuterVolumeSpecName: "kube-api-access-r4t6s") pod "6d990095-cc8f-479e-95a6-0c90ad349aab" (UID: "6d990095-cc8f-479e-95a6-0c90ad349aab"). InnerVolumeSpecName "kube-api-access-r4t6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.901696 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d990095-cc8f-479e-95a6-0c90ad349aab" (UID: "6d990095-cc8f-479e-95a6-0c90ad349aab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.943778 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.943815 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4t6s\" (UniqueName: \"kubernetes.io/projected/6d990095-cc8f-479e-95a6-0c90ad349aab-kube-api-access-r4t6s\") on node \"crc\" DevicePath \"\"" Oct 07 21:40:18 crc kubenswrapper[4926]: I1007 21:40:18.943833 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d990095-cc8f-479e-95a6-0c90ad349aab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.295171 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerID="db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a" exitCode=0 Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.295232 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jpxpq" event={"ID":"6d990095-cc8f-479e-95a6-0c90ad349aab","Type":"ContainerDied","Data":"db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a"} Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.295280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jpxpq" event={"ID":"6d990095-cc8f-479e-95a6-0c90ad349aab","Type":"ContainerDied","Data":"80a14711614f13023709f1045b7322f91500613a53a9a2809b5f92bd7117ce18"} Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.295295 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jpxpq" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.295303 4926 scope.go:117] "RemoveContainer" containerID="db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.346638 4926 scope.go:117] "RemoveContainer" containerID="762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.352058 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jpxpq"] Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.364011 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jpxpq"] Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.376043 4926 scope.go:117] "RemoveContainer" containerID="138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.423056 4926 scope.go:117] "RemoveContainer" containerID="db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a" Oct 07 21:40:19 crc kubenswrapper[4926]: E1007 21:40:19.423547 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a\": container with ID starting with db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a not found: ID does not exist" containerID="db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.423586 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a"} err="failed to get container status \"db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a\": rpc error: code = NotFound desc = could not find container \"db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a\": container with ID starting with db860a4120a639261c3380e71826706a05193612b16b305a47c616e46f41d31a not found: ID does not exist" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.423606 4926 scope.go:117] "RemoveContainer" containerID="762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac" Oct 07 21:40:19 crc kubenswrapper[4926]: E1007 21:40:19.424025 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac\": container with ID starting with 762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac not found: ID does not exist" containerID="762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.424055 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac"} err="failed to get container status \"762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac\": rpc error: code = NotFound desc = could not find container \"762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac\": container with ID starting with 762cb3601c6490f4cfaf7e39b4dfce783b4654777d14143e957c86cb5c8492ac not found: ID does not exist" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.424073 4926 scope.go:117] "RemoveContainer" containerID="138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933" Oct 07 21:40:19 crc kubenswrapper[4926]: E1007 21:40:19.424526 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933\": container with ID starting with 138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933 not found: ID does not exist" containerID="138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933" Oct 07 21:40:19 crc kubenswrapper[4926]: I1007 21:40:19.424554 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933"} err="failed to get container status \"138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933\": rpc error: code = NotFound desc = could not find container \"138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933\": container with ID starting with 138eb5044290f8d3fdd9adbbab7436003ef09ceb1a6d0c64fb508b3b2525a933 not found: ID does not exist" Oct 07 21:40:20 crc kubenswrapper[4926]: I1007 21:40:20.701788 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" path="/var/lib/kubelet/pods/6d990095-cc8f-479e-95a6-0c90ad349aab/volumes" Oct 07 21:40:33 crc kubenswrapper[4926]: I1007 21:40:33.209324 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:40:33 crc kubenswrapper[4926]: I1007 21:40:33.211735 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.022334 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c4bxq"] Oct 07 21:40:39 crc kubenswrapper[4926]: E1007 21:40:39.023561 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="extract-content" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.023587 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="extract-content" Oct 07 21:40:39 crc kubenswrapper[4926]: E1007 21:40:39.023604 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="extract-utilities" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.023610 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="extract-utilities" Oct 07 21:40:39 crc kubenswrapper[4926]: E1007 21:40:39.023661 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="registry-server" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.023667 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="registry-server" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.023929 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d990095-cc8f-479e-95a6-0c90ad349aab" containerName="registry-server" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.025542 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.027222 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz86p\" (UniqueName: \"kubernetes.io/projected/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-kube-api-access-xz86p\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.027284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-catalog-content\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.027321 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-utilities\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.044977 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4bxq"] Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.129617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz86p\" (UniqueName: \"kubernetes.io/projected/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-kube-api-access-xz86p\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.129827 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-catalog-content\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.130348 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-catalog-content\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.130405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-utilities\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.130774 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-utilities\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.150068 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz86p\" (UniqueName: \"kubernetes.io/projected/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-kube-api-access-xz86p\") pod \"redhat-operators-c4bxq\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.346289 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:39 crc kubenswrapper[4926]: I1007 21:40:39.713313 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4bxq"] Oct 07 21:40:40 crc kubenswrapper[4926]: I1007 21:40:40.560155 4926 generic.go:334] "Generic (PLEG): container finished" podID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerID="6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e" exitCode=0 Oct 07 21:40:40 crc kubenswrapper[4926]: I1007 21:40:40.560244 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4bxq" event={"ID":"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f","Type":"ContainerDied","Data":"6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e"} Oct 07 21:40:40 crc kubenswrapper[4926]: I1007 21:40:40.560447 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4bxq" event={"ID":"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f","Type":"ContainerStarted","Data":"9abc82652c20ebd0de3fd97d861bcb85a68cbe54a3b30a3b4142b8e6dfc31e6e"} Oct 07 21:40:42 crc kubenswrapper[4926]: I1007 21:40:42.583677 4926 generic.go:334] "Generic (PLEG): container finished" podID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerID="314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17" exitCode=0 Oct 07 21:40:42 crc kubenswrapper[4926]: I1007 21:40:42.583790 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4bxq" event={"ID":"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f","Type":"ContainerDied","Data":"314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17"} Oct 07 21:40:43 crc kubenswrapper[4926]: I1007 21:40:43.603901 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4bxq" event={"ID":"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f","Type":"ContainerStarted","Data":"3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083"} Oct 07 21:40:43 crc kubenswrapper[4926]: I1007 21:40:43.629923 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c4bxq" podStartSLOduration=2.166911973 podStartE2EDuration="4.629901206s" podCreationTimestamp="2025-10-07 21:40:39 +0000 UTC" firstStartedPulling="2025-10-07 21:40:40.561704876 +0000 UTC m=+2690.600286026" lastFinishedPulling="2025-10-07 21:40:43.024694089 +0000 UTC m=+2693.063275259" observedRunningTime="2025-10-07 21:40:43.623317045 +0000 UTC m=+2693.661898235" watchObservedRunningTime="2025-10-07 21:40:43.629901206 +0000 UTC m=+2693.668482366" Oct 07 21:40:49 crc kubenswrapper[4926]: I1007 21:40:49.347075 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:49 crc kubenswrapper[4926]: I1007 21:40:49.347942 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:49 crc kubenswrapper[4926]: I1007 21:40:49.423394 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:49 crc kubenswrapper[4926]: I1007 21:40:49.755084 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:49 crc kubenswrapper[4926]: I1007 21:40:49.819411 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c4bxq"] Oct 07 21:40:51 crc kubenswrapper[4926]: I1007 21:40:51.729833 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c4bxq" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="registry-server" containerID="cri-o://3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083" gracePeriod=2 Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.249628 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.425972 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-utilities\") pod \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.426025 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz86p\" (UniqueName: \"kubernetes.io/projected/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-kube-api-access-xz86p\") pod \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.426062 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-catalog-content\") pod \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\" (UID: \"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f\") " Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.427487 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-utilities" (OuterVolumeSpecName: "utilities") pod "ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" (UID: "ce7713b0-df0a-4ec6-b857-2ceb6b41a42f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.432644 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-kube-api-access-xz86p" (OuterVolumeSpecName: "kube-api-access-xz86p") pod "ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" (UID: "ce7713b0-df0a-4ec6-b857-2ceb6b41a42f"). InnerVolumeSpecName "kube-api-access-xz86p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.500103 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" (UID: "ce7713b0-df0a-4ec6-b857-2ceb6b41a42f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.530269 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.530303 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz86p\" (UniqueName: \"kubernetes.io/projected/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-kube-api-access-xz86p\") on node \"crc\" DevicePath \"\"" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.530315 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.739201 4926 generic.go:334] "Generic (PLEG): container finished" podID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerID="3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083" exitCode=0 Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.740296 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4bxq" event={"ID":"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f","Type":"ContainerDied","Data":"3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083"} Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.740336 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4bxq" event={"ID":"ce7713b0-df0a-4ec6-b857-2ceb6b41a42f","Type":"ContainerDied","Data":"9abc82652c20ebd0de3fd97d861bcb85a68cbe54a3b30a3b4142b8e6dfc31e6e"} Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.740358 4926 scope.go:117] "RemoveContainer" containerID="3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.740544 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4bxq" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.764002 4926 scope.go:117] "RemoveContainer" containerID="314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.770285 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c4bxq"] Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.779455 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c4bxq"] Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.792587 4926 scope.go:117] "RemoveContainer" containerID="6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.839789 4926 scope.go:117] "RemoveContainer" containerID="3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083" Oct 07 21:40:52 crc kubenswrapper[4926]: E1007 21:40:52.840211 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083\": container with ID starting with 3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083 not found: ID does not exist" containerID="3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.840248 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083"} err="failed to get container status \"3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083\": rpc error: code = NotFound desc = could not find container \"3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083\": container with ID starting with 3235c47154f771405860d43245825a4324d9430a3653e6e46391c9b626cf3083 not found: ID does not exist" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.840269 4926 scope.go:117] "RemoveContainer" containerID="314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17" Oct 07 21:40:52 crc kubenswrapper[4926]: E1007 21:40:52.840471 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17\": container with ID starting with 314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17 not found: ID does not exist" containerID="314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.840494 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17"} err="failed to get container status \"314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17\": rpc error: code = NotFound desc = could not find container \"314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17\": container with ID starting with 314e4623f18cec46cd2d1bf06a10c66d4d6010164847a43448a841dba6952e17 not found: ID does not exist" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.840508 4926 scope.go:117] "RemoveContainer" containerID="6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e" Oct 07 21:40:52 crc kubenswrapper[4926]: E1007 21:40:52.840672 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e\": container with ID starting with 6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e not found: ID does not exist" containerID="6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e" Oct 07 21:40:52 crc kubenswrapper[4926]: I1007 21:40:52.840694 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e"} err="failed to get container status \"6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e\": rpc error: code = NotFound desc = could not find container \"6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e\": container with ID starting with 6198c6ab16867d5ec53e317b4c974435460628e00587fb11e71d886a833b915e not found: ID does not exist" Oct 07 21:40:54 crc kubenswrapper[4926]: I1007 21:40:54.702750 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" path="/var/lib/kubelet/pods/ce7713b0-df0a-4ec6-b857-2ceb6b41a42f/volumes" Oct 07 21:41:03 crc kubenswrapper[4926]: I1007 21:41:03.209498 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:41:03 crc kubenswrapper[4926]: I1007 21:41:03.209935 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:41:09 crc kubenswrapper[4926]: I1007 21:41:09.988947 4926 generic.go:334] "Generic (PLEG): container finished" podID="0273b882-c15f-4ca8-b786-60fb42c205ab" containerID="fc12c5b185971822d0fa3570f5befe43c57940f0163b192eca577c66bbb59494" exitCode=0 Oct 07 21:41:09 crc kubenswrapper[4926]: I1007 21:41:09.989073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" event={"ID":"0273b882-c15f-4ca8-b786-60fb42c205ab","Type":"ContainerDied","Data":"fc12c5b185971822d0fa3570f5befe43c57940f0163b192eca577c66bbb59494"} Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.463733 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.660702 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-0\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.660949 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-combined-ca-bundle\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.661020 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-extra-config-0\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.661177 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-1\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.661517 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmczx\" (UniqueName: \"kubernetes.io/projected/0273b882-c15f-4ca8-b786-60fb42c205ab-kube-api-access-xmczx\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.661603 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-inventory\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.661760 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-0\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.662308 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-ssh-key\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.662395 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-1\") pod \"0273b882-c15f-4ca8-b786-60fb42c205ab\" (UID: \"0273b882-c15f-4ca8-b786-60fb42c205ab\") " Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.669650 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0273b882-c15f-4ca8-b786-60fb42c205ab-kube-api-access-xmczx" (OuterVolumeSpecName: "kube-api-access-xmczx") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "kube-api-access-xmczx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.670516 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.691502 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.697368 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.700033 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.701055 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-inventory" (OuterVolumeSpecName: "inventory") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.706373 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.714590 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.730327 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "0273b882-c15f-4ca8-b786-60fb42c205ab" (UID: "0273b882-c15f-4ca8-b786-60fb42c205ab"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766454 4926 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766504 4926 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766524 4926 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766545 4926 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766602 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmczx\" (UniqueName: \"kubernetes.io/projected/0273b882-c15f-4ca8-b786-60fb42c205ab-kube-api-access-xmczx\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766621 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766638 4926 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766654 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:11 crc kubenswrapper[4926]: I1007 21:41:11.766677 4926 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0273b882-c15f-4ca8-b786-60fb42c205ab-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.018044 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" event={"ID":"0273b882-c15f-4ca8-b786-60fb42c205ab","Type":"ContainerDied","Data":"be7ed188ead179f7ba7c656bf1f9863c3357003b7ec2e22385c09f56d6f4d35c"} Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.018506 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be7ed188ead179f7ba7c656bf1f9863c3357003b7ec2e22385c09f56d6f4d35c" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.018094 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t7hkb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.135781 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb"] Oct 07 21:41:12 crc kubenswrapper[4926]: E1007 21:41:12.136623 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="registry-server" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.136667 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="registry-server" Oct 07 21:41:12 crc kubenswrapper[4926]: E1007 21:41:12.136725 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0273b882-c15f-4ca8-b786-60fb42c205ab" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.136745 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0273b882-c15f-4ca8-b786-60fb42c205ab" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 07 21:41:12 crc kubenswrapper[4926]: E1007 21:41:12.136787 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="extract-content" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.136805 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="extract-content" Oct 07 21:41:12 crc kubenswrapper[4926]: E1007 21:41:12.136859 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="extract-utilities" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.136879 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="extract-utilities" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.137369 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7713b0-df0a-4ec6-b857-2ceb6b41a42f" containerName="registry-server" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.137427 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0273b882-c15f-4ca8-b786-60fb42c205ab" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.138590 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.141720 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.141915 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.142046 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.142157 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.142307 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ldlxc" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.149800 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb"] Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175291 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175482 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175533 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llb7l\" (UniqueName: \"kubernetes.io/projected/013ef09a-4927-4ca6-9beb-0537614a9187-kube-api-access-llb7l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175588 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175651 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175794 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.175893 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.278435 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.278627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.278772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.279836 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.279918 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llb7l\" (UniqueName: \"kubernetes.io/projected/013ef09a-4927-4ca6-9beb-0537614a9187-kube-api-access-llb7l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.280003 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.280091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.285543 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.286555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.286658 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.287084 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.287957 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.296952 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.299052 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llb7l\" (UniqueName: \"kubernetes.io/projected/013ef09a-4927-4ca6-9beb-0537614a9187-kube-api-access-llb7l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v99xb\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:12 crc kubenswrapper[4926]: I1007 21:41:12.464982 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:41:13 crc kubenswrapper[4926]: I1007 21:41:13.084182 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb"] Oct 07 21:41:14 crc kubenswrapper[4926]: I1007 21:41:14.040965 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" event={"ID":"013ef09a-4927-4ca6-9beb-0537614a9187","Type":"ContainerStarted","Data":"6f0d17529f0db85780ec4ad316bfc56257351b3ec229c1824177eb20745ae14b"} Oct 07 21:41:14 crc kubenswrapper[4926]: I1007 21:41:14.043399 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" event={"ID":"013ef09a-4927-4ca6-9beb-0537614a9187","Type":"ContainerStarted","Data":"6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d"} Oct 07 21:41:14 crc kubenswrapper[4926]: I1007 21:41:14.068742 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" podStartSLOduration=1.6013431900000001 podStartE2EDuration="2.068720149s" podCreationTimestamp="2025-10-07 21:41:12 +0000 UTC" firstStartedPulling="2025-10-07 21:41:13.089649766 +0000 UTC m=+2723.128230926" lastFinishedPulling="2025-10-07 21:41:13.557026725 +0000 UTC m=+2723.595607885" observedRunningTime="2025-10-07 21:41:14.065113165 +0000 UTC m=+2724.103694345" watchObservedRunningTime="2025-10-07 21:41:14.068720149 +0000 UTC m=+2724.107301309" Oct 07 21:41:33 crc kubenswrapper[4926]: I1007 21:41:33.209781 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:41:33 crc kubenswrapper[4926]: I1007 21:41:33.210526 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:41:33 crc kubenswrapper[4926]: I1007 21:41:33.210570 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:41:33 crc kubenswrapper[4926]: I1007 21:41:33.211332 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"45407472135a43dc033ae09daa12cedb15f24f7ec62fc31880605bb03ece9f39"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:41:33 crc kubenswrapper[4926]: I1007 21:41:33.211385 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://45407472135a43dc033ae09daa12cedb15f24f7ec62fc31880605bb03ece9f39" gracePeriod=600 Oct 07 21:41:34 crc kubenswrapper[4926]: I1007 21:41:34.270475 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="45407472135a43dc033ae09daa12cedb15f24f7ec62fc31880605bb03ece9f39" exitCode=0 Oct 07 21:41:34 crc kubenswrapper[4926]: I1007 21:41:34.270542 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"45407472135a43dc033ae09daa12cedb15f24f7ec62fc31880605bb03ece9f39"} Oct 07 21:41:34 crc kubenswrapper[4926]: I1007 21:41:34.271135 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2"} Oct 07 21:41:34 crc kubenswrapper[4926]: I1007 21:41:34.271162 4926 scope.go:117] "RemoveContainer" containerID="0461d24cb6a3d054b5789cbd520024d1e2e4f2bb97347cd273b3e344d8e9a7c8" Oct 07 21:43:33 crc kubenswrapper[4926]: I1007 21:43:33.210175 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:43:33 crc kubenswrapper[4926]: I1007 21:43:33.210793 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:43:48 crc kubenswrapper[4926]: I1007 21:43:48.845688 4926 generic.go:334] "Generic (PLEG): container finished" podID="013ef09a-4927-4ca6-9beb-0537614a9187" containerID="6f0d17529f0db85780ec4ad316bfc56257351b3ec229c1824177eb20745ae14b" exitCode=0 Oct 07 21:43:48 crc kubenswrapper[4926]: I1007 21:43:48.846435 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" event={"ID":"013ef09a-4927-4ca6-9beb-0537614a9187","Type":"ContainerDied","Data":"6f0d17529f0db85780ec4ad316bfc56257351b3ec229c1824177eb20745ae14b"} Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.358893 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469329 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-1\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469489 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-0\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469620 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ssh-key\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469693 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-2\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469881 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-telemetry-combined-ca-bundle\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-inventory\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.469948 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llb7l\" (UniqueName: \"kubernetes.io/projected/013ef09a-4927-4ca6-9beb-0537614a9187-kube-api-access-llb7l\") pod \"013ef09a-4927-4ca6-9beb-0537614a9187\" (UID: \"013ef09a-4927-4ca6-9beb-0537614a9187\") " Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.480747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/013ef09a-4927-4ca6-9beb-0537614a9187-kube-api-access-llb7l" (OuterVolumeSpecName: "kube-api-access-llb7l") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "kube-api-access-llb7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.483832 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.506931 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.507180 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.519184 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-inventory" (OuterVolumeSpecName: "inventory") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.524692 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.531360 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "013ef09a-4927-4ca6-9beb-0537614a9187" (UID: "013ef09a-4927-4ca6-9beb-0537614a9187"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572712 4926 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572762 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572783 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llb7l\" (UniqueName: \"kubernetes.io/projected/013ef09a-4927-4ca6-9beb-0537614a9187-kube-api-access-llb7l\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572802 4926 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572822 4926 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572839 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.572859 4926 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/013ef09a-4927-4ca6-9beb-0537614a9187-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.876557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" event={"ID":"013ef09a-4927-4ca6-9beb-0537614a9187","Type":"ContainerDied","Data":"6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d"} Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.876605 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d" Oct 07 21:43:50 crc kubenswrapper[4926]: I1007 21:43:50.876665 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v99xb" Oct 07 21:44:00 crc kubenswrapper[4926]: E1007 21:44:00.457313 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice/crio-6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice\": RecentStats: unable to find data in memory cache]" Oct 07 21:44:03 crc kubenswrapper[4926]: I1007 21:44:03.209468 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:44:03 crc kubenswrapper[4926]: I1007 21:44:03.209847 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:44:10 crc kubenswrapper[4926]: E1007 21:44:10.790289 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice/crio-6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d\": RecentStats: unable to find data in memory cache]" Oct 07 21:44:21 crc kubenswrapper[4926]: E1007 21:44:21.122750 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice/crio-6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice\": RecentStats: unable to find data in memory cache]" Oct 07 21:44:26 crc kubenswrapper[4926]: I1007 21:44:26.755514 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:44:26 crc kubenswrapper[4926]: I1007 21:44:26.756395 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="prometheus" containerID="cri-o://94424ffca67f31b70557f548664110fc9d36d2d1072473eeb4975b1fe157e99a" gracePeriod=600 Oct 07 21:44:26 crc kubenswrapper[4926]: I1007 21:44:26.756474 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="config-reloader" containerID="cri-o://a5905b1793ca91f9fcf57228ef018d794db4a4aa6dd1c3db25c3c2e0a0fd468e" gracePeriod=600 Oct 07 21:44:26 crc kubenswrapper[4926]: I1007 21:44:26.756498 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="thanos-sidecar" containerID="cri-o://e15617ef19af5594b54256816f4c297916f3694e1c8cf532720ac567c2db4d79" gracePeriod=600 Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.336093 4926 generic.go:334] "Generic (PLEG): container finished" podID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerID="e15617ef19af5594b54256816f4c297916f3694e1c8cf532720ac567c2db4d79" exitCode=0 Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.336333 4926 generic.go:334] "Generic (PLEG): container finished" podID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerID="a5905b1793ca91f9fcf57228ef018d794db4a4aa6dd1c3db25c3c2e0a0fd468e" exitCode=0 Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.336341 4926 generic.go:334] "Generic (PLEG): container finished" podID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerID="94424ffca67f31b70557f548664110fc9d36d2d1072473eeb4975b1fe157e99a" exitCode=0 Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.336165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerDied","Data":"e15617ef19af5594b54256816f4c297916f3694e1c8cf532720ac567c2db4d79"} Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.336378 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerDied","Data":"a5905b1793ca91f9fcf57228ef018d794db4a4aa6dd1c3db25c3c2e0a0fd468e"} Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.336394 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerDied","Data":"94424ffca67f31b70557f548664110fc9d36d2d1072473eeb4975b1fe157e99a"} Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.654574 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.735317 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7db60520-2468-44b2-9ca3-99c1bfe04bad-prometheus-metric-storage-rulefiles-0\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.735392 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nllkk\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-kube-api-access-nllkk\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.735432 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7db60520-2468-44b2-9ca3-99c1bfe04bad-config-out\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.735478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-secret-combined-ca-bundle\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.735552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736285 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db60520-2468-44b2-9ca3-99c1bfe04bad-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-config\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736606 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736636 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736672 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736694 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-thanos-prometheus-http-client-file\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.736828 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-tls-assets\") pod \"7db60520-2468-44b2-9ca3-99c1bfe04bad\" (UID: \"7db60520-2468-44b2-9ca3-99c1bfe04bad\") " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.745022 4926 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7db60520-2468-44b2-9ca3-99c1bfe04bad-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.745674 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-kube-api-access-nllkk" (OuterVolumeSpecName: "kube-api-access-nllkk") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "kube-api-access-nllkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.745833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.745877 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7db60520-2468-44b2-9ca3-99c1bfe04bad-config-out" (OuterVolumeSpecName: "config-out") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.746021 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.748058 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.753801 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-config" (OuterVolumeSpecName: "config") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.762164 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.764966 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.775993 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "pvc-25bd2660-6460-4b17-9118-3d7559e7645c". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846807 4926 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846856 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nllkk\" (UniqueName: \"kubernetes.io/projected/7db60520-2468-44b2-9ca3-99c1bfe04bad-kube-api-access-nllkk\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846871 4926 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7db60520-2468-44b2-9ca3-99c1bfe04bad-config-out\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846884 4926 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846896 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846936 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") on node \"crc\" " Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846952 4926 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846967 4926 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.846981 4926 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.864623 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config" (OuterVolumeSpecName: "web-config") pod "7db60520-2468-44b2-9ca3-99c1bfe04bad" (UID: "7db60520-2468-44b2-9ca3-99c1bfe04bad"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.876817 4926 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.877011 4926 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-25bd2660-6460-4b17-9118-3d7559e7645c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c") on node "crc" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.949667 4926 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7db60520-2468-44b2-9ca3-99c1bfe04bad-web-config\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:27 crc kubenswrapper[4926]: I1007 21:44:27.949702 4926 reconciler_common.go:293] "Volume detached for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") on node \"crc\" DevicePath \"\"" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.355894 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7db60520-2468-44b2-9ca3-99c1bfe04bad","Type":"ContainerDied","Data":"4a1eb84374483306bb63ca335c378799246b25a506fd0f62948d8a5c0574aa57"} Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.357811 4926 scope.go:117] "RemoveContainer" containerID="e15617ef19af5594b54256816f4c297916f3694e1c8cf532720ac567c2db4d79" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.358079 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.385813 4926 scope.go:117] "RemoveContainer" containerID="a5905b1793ca91f9fcf57228ef018d794db4a4aa6dd1c3db25c3c2e0a0fd468e" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.423451 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.440778 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.446145 4926 scope.go:117] "RemoveContainer" containerID="94424ffca67f31b70557f548664110fc9d36d2d1072473eeb4975b1fe157e99a" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484090 4926 scope.go:117] "RemoveContainer" containerID="cef19952d2ea95e04fd898fa6be4266495564473841e4e2300f410ecd31b8aa0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484239 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:44:28 crc kubenswrapper[4926]: E1007 21:44:28.484642 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013ef09a-4927-4ca6-9beb-0537614a9187" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484658 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="013ef09a-4927-4ca6-9beb-0537614a9187" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 07 21:44:28 crc kubenswrapper[4926]: E1007 21:44:28.484679 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="thanos-sidecar" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484685 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="thanos-sidecar" Oct 07 21:44:28 crc kubenswrapper[4926]: E1007 21:44:28.484702 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="prometheus" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484708 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="prometheus" Oct 07 21:44:28 crc kubenswrapper[4926]: E1007 21:44:28.484723 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="config-reloader" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484730 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="config-reloader" Oct 07 21:44:28 crc kubenswrapper[4926]: E1007 21:44:28.484744 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="init-config-reloader" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484753 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="init-config-reloader" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484948 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="thanos-sidecar" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484971 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="prometheus" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484979 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="013ef09a-4927-4ca6-9beb-0537614a9187" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.484987 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" containerName="config-reloader" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.487024 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.494699 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.494804 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.494984 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.495000 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hckhp" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.495079 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.499636 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.517939 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564410 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/47ab84ef-8080-42f6-9467-6d1439275f65-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564495 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564528 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564606 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/47ab84ef-8080-42f6-9467-6d1439275f65-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564627 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564817 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-config\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564861 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/47ab84ef-8080-42f6-9467-6d1439275f65-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564899 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.564979 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m7vf\" (UniqueName: \"kubernetes.io/projected/47ab84ef-8080-42f6-9467-6d1439275f65-kube-api-access-6m7vf\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.565009 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.666821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.666901 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m7vf\" (UniqueName: \"kubernetes.io/projected/47ab84ef-8080-42f6-9467-6d1439275f65-kube-api-access-6m7vf\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.666936 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.666960 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.666984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/47ab84ef-8080-42f6-9467-6d1439275f65-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.667033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.667061 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.667133 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/47ab84ef-8080-42f6-9467-6d1439275f65-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.667152 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.667174 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-config\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.667244 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/47ab84ef-8080-42f6-9467-6d1439275f65-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.668605 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/47ab84ef-8080-42f6-9467-6d1439275f65-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.670415 4926 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.670457 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/981924a86d194783a644e1657994993df571041fb46e6d3f243affb52a3a59ba/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.672339 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.675791 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.676126 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/47ab84ef-8080-42f6-9467-6d1439275f65-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.676366 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.676745 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.676775 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/47ab84ef-8080-42f6-9467-6d1439275f65-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.677907 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-config\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.681525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/47ab84ef-8080-42f6-9467-6d1439275f65-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.691581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m7vf\" (UniqueName: \"kubernetes.io/projected/47ab84ef-8080-42f6-9467-6d1439275f65-kube-api-access-6m7vf\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.692688 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db60520-2468-44b2-9ca3-99c1bfe04bad" path="/var/lib/kubelet/pods/7db60520-2468-44b2-9ca3-99c1bfe04bad/volumes" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.723268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25bd2660-6460-4b17-9118-3d7559e7645c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25bd2660-6460-4b17-9118-3d7559e7645c\") pod \"prometheus-metric-storage-0\" (UID: \"47ab84ef-8080-42f6-9467-6d1439275f65\") " pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:28 crc kubenswrapper[4926]: I1007 21:44:28.905602 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:29 crc kubenswrapper[4926]: I1007 21:44:29.423307 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 21:44:30 crc kubenswrapper[4926]: I1007 21:44:30.392356 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"47ab84ef-8080-42f6-9467-6d1439275f65","Type":"ContainerStarted","Data":"4f567a465b9fef46fec1cd07d2d4dfff0a4625d4b77745c4d19a110a8e3808fe"} Oct 07 21:44:31 crc kubenswrapper[4926]: E1007 21:44:31.415962 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice/crio-6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice\": RecentStats: unable to find data in memory cache]" Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.209379 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.209855 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.209937 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.211348 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.211454 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" gracePeriod=600 Oct 07 21:44:33 crc kubenswrapper[4926]: E1007 21:44:33.342715 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.426774 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" exitCode=0 Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.426871 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2"} Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.427086 4926 scope.go:117] "RemoveContainer" containerID="45407472135a43dc033ae09daa12cedb15f24f7ec62fc31880605bb03ece9f39" Oct 07 21:44:33 crc kubenswrapper[4926]: I1007 21:44:33.427819 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:44:33 crc kubenswrapper[4926]: E1007 21:44:33.428131 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:44:34 crc kubenswrapper[4926]: I1007 21:44:34.442736 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"47ab84ef-8080-42f6-9467-6d1439275f65","Type":"ContainerStarted","Data":"4eff6ea411fa723e2a7117f5f1c2ff73e55b4e5c2e464b969267499f2ac8be4b"} Oct 07 21:44:41 crc kubenswrapper[4926]: E1007 21:44:41.735881 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013ef09a_4927_4ca6_9beb_0537614a9187.slice/crio-6649421a6ae8c7c00b45447b77ca07ac1330066ffee57fee250fa871829a581d\": RecentStats: unable to find data in memory cache]" Oct 07 21:44:45 crc kubenswrapper[4926]: I1007 21:44:45.567250 4926 generic.go:334] "Generic (PLEG): container finished" podID="47ab84ef-8080-42f6-9467-6d1439275f65" containerID="4eff6ea411fa723e2a7117f5f1c2ff73e55b4e5c2e464b969267499f2ac8be4b" exitCode=0 Oct 07 21:44:45 crc kubenswrapper[4926]: I1007 21:44:45.567319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"47ab84ef-8080-42f6-9467-6d1439275f65","Type":"ContainerDied","Data":"4eff6ea411fa723e2a7117f5f1c2ff73e55b4e5c2e464b969267499f2ac8be4b"} Oct 07 21:44:46 crc kubenswrapper[4926]: I1007 21:44:46.580145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"47ab84ef-8080-42f6-9467-6d1439275f65","Type":"ContainerStarted","Data":"2b68eee76e3c97cc5be8abd651a6b124eadf138b23e5040c115359390f4b0549"} Oct 07 21:44:47 crc kubenswrapper[4926]: I1007 21:44:47.680079 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:44:47 crc kubenswrapper[4926]: E1007 21:44:47.680714 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:44:50 crc kubenswrapper[4926]: I1007 21:44:50.631148 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"47ab84ef-8080-42f6-9467-6d1439275f65","Type":"ContainerStarted","Data":"7096ee581384410087609526674d40470695b301e6ef95b17ebc163ce6878d4c"} Oct 07 21:44:50 crc kubenswrapper[4926]: I1007 21:44:50.632030 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"47ab84ef-8080-42f6-9467-6d1439275f65","Type":"ContainerStarted","Data":"0d2ab4920109198281fade255b6da2b2703de9a77d3354d376a1aa5489e59093"} Oct 07 21:44:50 crc kubenswrapper[4926]: I1007 21:44:50.679650 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=22.679630331 podStartE2EDuration="22.679630331s" podCreationTimestamp="2025-10-07 21:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 21:44:50.665909573 +0000 UTC m=+2940.704490743" watchObservedRunningTime="2025-10-07 21:44:50.679630331 +0000 UTC m=+2940.718211491" Oct 07 21:44:53 crc kubenswrapper[4926]: I1007 21:44:53.906690 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:58 crc kubenswrapper[4926]: I1007 21:44:58.907737 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:58 crc kubenswrapper[4926]: I1007 21:44:58.914489 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 07 21:44:59 crc kubenswrapper[4926]: I1007 21:44:59.771257 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.184948 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j"] Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.189284 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.193779 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.194046 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.217334 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j"] Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.356241 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833ec329-91dd-43e2-a699-7857055ead36-secret-volume\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.356307 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833ec329-91dd-43e2-a699-7857055ead36-config-volume\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.356341 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x6r9\" (UniqueName: \"kubernetes.io/projected/833ec329-91dd-43e2-a699-7857055ead36-kube-api-access-7x6r9\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.457835 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833ec329-91dd-43e2-a699-7857055ead36-secret-volume\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.458111 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833ec329-91dd-43e2-a699-7857055ead36-config-volume\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.458221 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x6r9\" (UniqueName: \"kubernetes.io/projected/833ec329-91dd-43e2-a699-7857055ead36-kube-api-access-7x6r9\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.459303 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833ec329-91dd-43e2-a699-7857055ead36-config-volume\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.473309 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833ec329-91dd-43e2-a699-7857055ead36-secret-volume\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.476856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x6r9\" (UniqueName: \"kubernetes.io/projected/833ec329-91dd-43e2-a699-7857055ead36-kube-api-access-7x6r9\") pod \"collect-profiles-29331225-pkr6j\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.516922 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.688693 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:45:00 crc kubenswrapper[4926]: E1007 21:45:00.689507 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:45:00 crc kubenswrapper[4926]: I1007 21:45:00.992521 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j"] Oct 07 21:45:01 crc kubenswrapper[4926]: W1007 21:45:01.000059 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod833ec329_91dd_43e2_a699_7857055ead36.slice/crio-b727e317df0f35f9a20f95fc0a33ae60c5834aa4a5b025af3f73d3545b4f8bd8 WatchSource:0}: Error finding container b727e317df0f35f9a20f95fc0a33ae60c5834aa4a5b025af3f73d3545b4f8bd8: Status 404 returned error can't find the container with id b727e317df0f35f9a20f95fc0a33ae60c5834aa4a5b025af3f73d3545b4f8bd8 Oct 07 21:45:01 crc kubenswrapper[4926]: I1007 21:45:01.838964 4926 generic.go:334] "Generic (PLEG): container finished" podID="833ec329-91dd-43e2-a699-7857055ead36" containerID="539eae27cc239789eb0de7e5b7b6cea4b4234136945e47a0d656677380e22dee" exitCode=0 Oct 07 21:45:01 crc kubenswrapper[4926]: I1007 21:45:01.839323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" event={"ID":"833ec329-91dd-43e2-a699-7857055ead36","Type":"ContainerDied","Data":"539eae27cc239789eb0de7e5b7b6cea4b4234136945e47a0d656677380e22dee"} Oct 07 21:45:01 crc kubenswrapper[4926]: I1007 21:45:01.839366 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" event={"ID":"833ec329-91dd-43e2-a699-7857055ead36","Type":"ContainerStarted","Data":"b727e317df0f35f9a20f95fc0a33ae60c5834aa4a5b025af3f73d3545b4f8bd8"} Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.278611 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.336721 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x6r9\" (UniqueName: \"kubernetes.io/projected/833ec329-91dd-43e2-a699-7857055ead36-kube-api-access-7x6r9\") pod \"833ec329-91dd-43e2-a699-7857055ead36\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.337317 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833ec329-91dd-43e2-a699-7857055ead36-config-volume\") pod \"833ec329-91dd-43e2-a699-7857055ead36\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.337433 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833ec329-91dd-43e2-a699-7857055ead36-secret-volume\") pod \"833ec329-91dd-43e2-a699-7857055ead36\" (UID: \"833ec329-91dd-43e2-a699-7857055ead36\") " Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.337963 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/833ec329-91dd-43e2-a699-7857055ead36-config-volume" (OuterVolumeSpecName: "config-volume") pod "833ec329-91dd-43e2-a699-7857055ead36" (UID: "833ec329-91dd-43e2-a699-7857055ead36"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.338414 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833ec329-91dd-43e2-a699-7857055ead36-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.346655 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833ec329-91dd-43e2-a699-7857055ead36-kube-api-access-7x6r9" (OuterVolumeSpecName: "kube-api-access-7x6r9") pod "833ec329-91dd-43e2-a699-7857055ead36" (UID: "833ec329-91dd-43e2-a699-7857055ead36"). InnerVolumeSpecName "kube-api-access-7x6r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.348518 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/833ec329-91dd-43e2-a699-7857055ead36-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "833ec329-91dd-43e2-a699-7857055ead36" (UID: "833ec329-91dd-43e2-a699-7857055ead36"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.439582 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x6r9\" (UniqueName: \"kubernetes.io/projected/833ec329-91dd-43e2-a699-7857055ead36-kube-api-access-7x6r9\") on node \"crc\" DevicePath \"\"" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.439621 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833ec329-91dd-43e2-a699-7857055ead36-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.873274 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" event={"ID":"833ec329-91dd-43e2-a699-7857055ead36","Type":"ContainerDied","Data":"b727e317df0f35f9a20f95fc0a33ae60c5834aa4a5b025af3f73d3545b4f8bd8"} Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.873335 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b727e317df0f35f9a20f95fc0a33ae60c5834aa4a5b025af3f73d3545b4f8bd8" Oct 07 21:45:03 crc kubenswrapper[4926]: I1007 21:45:03.873416 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j" Oct 07 21:45:04 crc kubenswrapper[4926]: I1007 21:45:04.368278 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v"] Oct 07 21:45:04 crc kubenswrapper[4926]: I1007 21:45:04.380648 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331180-48f9v"] Oct 07 21:45:04 crc kubenswrapper[4926]: I1007 21:45:04.702059 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43588f58-5468-4b85-8678-58bdd40823c9" path="/var/lib/kubelet/pods/43588f58-5468-4b85-8678-58bdd40823c9/volumes" Oct 07 21:45:05 crc kubenswrapper[4926]: I1007 21:45:05.353761 4926 scope.go:117] "RemoveContainer" containerID="6f7dea97f89e91b35d5d482bb6bba6f0beb1d46f20359e08a6a5ae4fc519776f" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.713291 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bb2w6"] Oct 07 21:45:10 crc kubenswrapper[4926]: E1007 21:45:10.714983 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833ec329-91dd-43e2-a699-7857055ead36" containerName="collect-profiles" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.715017 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="833ec329-91dd-43e2-a699-7857055ead36" containerName="collect-profiles" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.715591 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="833ec329-91dd-43e2-a699-7857055ead36" containerName="collect-profiles" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.719137 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.729474 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bb2w6"] Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.817141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-utilities\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.817559 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9nwj\" (UniqueName: \"kubernetes.io/projected/eb54d8f0-dc7e-4805-b484-5311e096ec76-kube-api-access-b9nwj\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.817609 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-catalog-content\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.919785 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-catalog-content\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.919932 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-utilities\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.920016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9nwj\" (UniqueName: \"kubernetes.io/projected/eb54d8f0-dc7e-4805-b484-5311e096ec76-kube-api-access-b9nwj\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.920310 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-catalog-content\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.920551 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-utilities\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:10 crc kubenswrapper[4926]: I1007 21:45:10.940152 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9nwj\" (UniqueName: \"kubernetes.io/projected/eb54d8f0-dc7e-4805-b484-5311e096ec76-kube-api-access-b9nwj\") pod \"community-operators-bb2w6\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.046709 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.578767 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bb2w6"] Oct 07 21:45:11 crc kubenswrapper[4926]: W1007 21:45:11.584123 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb54d8f0_dc7e_4805_b484_5311e096ec76.slice/crio-8887e3534f8d1b4a68e13e16c61ffd3f91815d953e01934c439e6672a0961390 WatchSource:0}: Error finding container 8887e3534f8d1b4a68e13e16c61ffd3f91815d953e01934c439e6672a0961390: Status 404 returned error can't find the container with id 8887e3534f8d1b4a68e13e16c61ffd3f91815d953e01934c439e6672a0961390 Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.680133 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:45:11 crc kubenswrapper[4926]: E1007 21:45:11.680621 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.969109 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerID="cb0f3778c926c95517b9ad66a4481f8093f26c2d9c98740ec25a67b799523654" exitCode=0 Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.969262 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerDied","Data":"cb0f3778c926c95517b9ad66a4481f8093f26c2d9c98740ec25a67b799523654"} Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.969451 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerStarted","Data":"8887e3534f8d1b4a68e13e16c61ffd3f91815d953e01934c439e6672a0961390"} Oct 07 21:45:11 crc kubenswrapper[4926]: I1007 21:45:11.972169 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:45:12 crc kubenswrapper[4926]: I1007 21:45:12.984907 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerStarted","Data":"0cced673fed1741cd715610dca4d89b5282070d85f5792adb847d9d71b12da0f"} Oct 07 21:45:14 crc kubenswrapper[4926]: I1007 21:45:13.999678 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerID="0cced673fed1741cd715610dca4d89b5282070d85f5792adb847d9d71b12da0f" exitCode=0 Oct 07 21:45:14 crc kubenswrapper[4926]: I1007 21:45:13.999765 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerDied","Data":"0cced673fed1741cd715610dca4d89b5282070d85f5792adb847d9d71b12da0f"} Oct 07 21:45:15 crc kubenswrapper[4926]: I1007 21:45:15.013258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerStarted","Data":"38b3c036c7b30e1b3250f7597ee4ef0a1fe3d2b6d83a9a069dc84745e04f0e26"} Oct 07 21:45:15 crc kubenswrapper[4926]: I1007 21:45:15.052523 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bb2w6" podStartSLOduration=2.51051692 podStartE2EDuration="5.052493335s" podCreationTimestamp="2025-10-07 21:45:10 +0000 UTC" firstStartedPulling="2025-10-07 21:45:11.971857893 +0000 UTC m=+2962.010439053" lastFinishedPulling="2025-10-07 21:45:14.513834278 +0000 UTC m=+2964.552415468" observedRunningTime="2025-10-07 21:45:15.035936005 +0000 UTC m=+2965.074517195" watchObservedRunningTime="2025-10-07 21:45:15.052493335 +0000 UTC m=+2965.091074525" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.137213 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.139168 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.142349 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-86k4k" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.143185 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.143410 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.143521 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.157295 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.302949 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303061 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303099 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-config-data\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303124 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn6h8\" (UniqueName: \"kubernetes.io/projected/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-kube-api-access-bn6h8\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303170 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303188 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303286 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303503 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.303568 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.405983 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.406339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-config-data\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.406517 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn6h8\" (UniqueName: \"kubernetes.io/projected/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-kube-api-access-bn6h8\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.406707 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.406857 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.407032 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.407168 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.407333 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.407535 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.407739 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.406954 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.407726 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-config-data\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.408394 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.408785 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.412872 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.412889 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.419610 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.426935 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn6h8\" (UniqueName: \"kubernetes.io/projected/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-kube-api-access-bn6h8\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.439394 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.496292 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 07 21:45:19 crc kubenswrapper[4926]: I1007 21:45:19.989535 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 07 21:45:19 crc kubenswrapper[4926]: W1007 21:45:19.993818 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25a6326c_d6df_46b2_b646_0c37aa3bb0ed.slice/crio-2f5ce6af17884b917be1944f264e21cd76f0f4e35b1f8c0d9f7040733be60fec WatchSource:0}: Error finding container 2f5ce6af17884b917be1944f264e21cd76f0f4e35b1f8c0d9f7040733be60fec: Status 404 returned error can't find the container with id 2f5ce6af17884b917be1944f264e21cd76f0f4e35b1f8c0d9f7040733be60fec Oct 07 21:45:20 crc kubenswrapper[4926]: I1007 21:45:20.067817 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"25a6326c-d6df-46b2-b646-0c37aa3bb0ed","Type":"ContainerStarted","Data":"2f5ce6af17884b917be1944f264e21cd76f0f4e35b1f8c0d9f7040733be60fec"} Oct 07 21:45:21 crc kubenswrapper[4926]: I1007 21:45:21.047049 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:21 crc kubenswrapper[4926]: I1007 21:45:21.047159 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:21 crc kubenswrapper[4926]: I1007 21:45:21.125248 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:22 crc kubenswrapper[4926]: I1007 21:45:22.138489 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:22 crc kubenswrapper[4926]: I1007 21:45:22.187840 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bb2w6"] Oct 07 21:45:24 crc kubenswrapper[4926]: I1007 21:45:24.106242 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bb2w6" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="registry-server" containerID="cri-o://38b3c036c7b30e1b3250f7597ee4ef0a1fe3d2b6d83a9a069dc84745e04f0e26" gracePeriod=2 Oct 07 21:45:25 crc kubenswrapper[4926]: I1007 21:45:25.119707 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerID="38b3c036c7b30e1b3250f7597ee4ef0a1fe3d2b6d83a9a069dc84745e04f0e26" exitCode=0 Oct 07 21:45:25 crc kubenswrapper[4926]: I1007 21:45:25.119754 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerDied","Data":"38b3c036c7b30e1b3250f7597ee4ef0a1fe3d2b6d83a9a069dc84745e04f0e26"} Oct 07 21:45:25 crc kubenswrapper[4926]: I1007 21:45:25.679351 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:45:25 crc kubenswrapper[4926]: E1007 21:45:25.679618 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.306163 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.445696 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-catalog-content\") pod \"eb54d8f0-dc7e-4805-b484-5311e096ec76\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.445764 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-utilities\") pod \"eb54d8f0-dc7e-4805-b484-5311e096ec76\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.445867 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9nwj\" (UniqueName: \"kubernetes.io/projected/eb54d8f0-dc7e-4805-b484-5311e096ec76-kube-api-access-b9nwj\") pod \"eb54d8f0-dc7e-4805-b484-5311e096ec76\" (UID: \"eb54d8f0-dc7e-4805-b484-5311e096ec76\") " Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.446753 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-utilities" (OuterVolumeSpecName: "utilities") pod "eb54d8f0-dc7e-4805-b484-5311e096ec76" (UID: "eb54d8f0-dc7e-4805-b484-5311e096ec76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.449395 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb54d8f0-dc7e-4805-b484-5311e096ec76-kube-api-access-b9nwj" (OuterVolumeSpecName: "kube-api-access-b9nwj") pod "eb54d8f0-dc7e-4805-b484-5311e096ec76" (UID: "eb54d8f0-dc7e-4805-b484-5311e096ec76"). InnerVolumeSpecName "kube-api-access-b9nwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.511087 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb54d8f0-dc7e-4805-b484-5311e096ec76" (UID: "eb54d8f0-dc7e-4805-b484-5311e096ec76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.548398 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.548446 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d8f0-dc7e-4805-b484-5311e096ec76-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:45:29 crc kubenswrapper[4926]: I1007 21:45:29.548460 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9nwj\" (UniqueName: \"kubernetes.io/projected/eb54d8f0-dc7e-4805-b484-5311e096ec76-kube-api-access-b9nwj\") on node \"crc\" DevicePath \"\"" Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.179911 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bb2w6" Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.179909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bb2w6" event={"ID":"eb54d8f0-dc7e-4805-b484-5311e096ec76","Type":"ContainerDied","Data":"8887e3534f8d1b4a68e13e16c61ffd3f91815d953e01934c439e6672a0961390"} Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.180390 4926 scope.go:117] "RemoveContainer" containerID="38b3c036c7b30e1b3250f7597ee4ef0a1fe3d2b6d83a9a069dc84745e04f0e26" Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.185865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"25a6326c-d6df-46b2-b646-0c37aa3bb0ed","Type":"ContainerStarted","Data":"01e8104ab7c378d2406efefd870eca413f48000e6da4ed5034105b83502af529"} Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.219665 4926 scope.go:117] "RemoveContainer" containerID="0cced673fed1741cd715610dca4d89b5282070d85f5792adb847d9d71b12da0f" Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.221741 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.21149786 podStartE2EDuration="12.221714444s" podCreationTimestamp="2025-10-07 21:45:18 +0000 UTC" firstStartedPulling="2025-10-07 21:45:19.997015338 +0000 UTC m=+2970.035596528" lastFinishedPulling="2025-10-07 21:45:29.007231952 +0000 UTC m=+2979.045813112" observedRunningTime="2025-10-07 21:45:30.214923667 +0000 UTC m=+2980.253504837" watchObservedRunningTime="2025-10-07 21:45:30.221714444 +0000 UTC m=+2980.260295634" Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.252793 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bb2w6"] Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.255484 4926 scope.go:117] "RemoveContainer" containerID="cb0f3778c926c95517b9ad66a4481f8093f26c2d9c98740ec25a67b799523654" Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.266847 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bb2w6"] Oct 07 21:45:30 crc kubenswrapper[4926]: I1007 21:45:30.701385 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" path="/var/lib/kubelet/pods/eb54d8f0-dc7e-4805-b484-5311e096ec76/volumes" Oct 07 21:45:37 crc kubenswrapper[4926]: I1007 21:45:37.679673 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:45:37 crc kubenswrapper[4926]: E1007 21:45:37.681782 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:45:51 crc kubenswrapper[4926]: I1007 21:45:51.679922 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:45:51 crc kubenswrapper[4926]: E1007 21:45:51.681289 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.802955 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-svwlj"] Oct 07 21:45:58 crc kubenswrapper[4926]: E1007 21:45:58.804105 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="registry-server" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.804127 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="registry-server" Oct 07 21:45:58 crc kubenswrapper[4926]: E1007 21:45:58.804172 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="extract-utilities" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.804182 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="extract-utilities" Oct 07 21:45:58 crc kubenswrapper[4926]: E1007 21:45:58.804311 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="extract-content" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.804321 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="extract-content" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.804581 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb54d8f0-dc7e-4805-b484-5311e096ec76" containerName="registry-server" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.806881 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.823417 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-svwlj"] Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.885065 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8nll\" (UniqueName: \"kubernetes.io/projected/b72e553d-26ac-460e-94da-a5ffc59f9a8a-kube-api-access-j8nll\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.885184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-utilities\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.885363 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-catalog-content\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.987670 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8nll\" (UniqueName: \"kubernetes.io/projected/b72e553d-26ac-460e-94da-a5ffc59f9a8a-kube-api-access-j8nll\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.987754 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-utilities\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.987864 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-catalog-content\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.988311 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-catalog-content\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:58 crc kubenswrapper[4926]: I1007 21:45:58.988799 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-utilities\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:59 crc kubenswrapper[4926]: I1007 21:45:59.009936 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8nll\" (UniqueName: \"kubernetes.io/projected/b72e553d-26ac-460e-94da-a5ffc59f9a8a-kube-api-access-j8nll\") pod \"redhat-marketplace-svwlj\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:59 crc kubenswrapper[4926]: I1007 21:45:59.150905 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:45:59 crc kubenswrapper[4926]: I1007 21:45:59.639163 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-svwlj"] Oct 07 21:46:00 crc kubenswrapper[4926]: I1007 21:46:00.561823 4926 generic.go:334] "Generic (PLEG): container finished" podID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerID="5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50" exitCode=0 Oct 07 21:46:00 crc kubenswrapper[4926]: I1007 21:46:00.561924 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerDied","Data":"5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50"} Oct 07 21:46:00 crc kubenswrapper[4926]: I1007 21:46:00.562169 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerStarted","Data":"48596de64612e8cddc9920884858855d7953683b4768e002bf9b773c0a17c22d"} Oct 07 21:46:01 crc kubenswrapper[4926]: I1007 21:46:01.578495 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerStarted","Data":"5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48"} Oct 07 21:46:02 crc kubenswrapper[4926]: I1007 21:46:02.595790 4926 generic.go:334] "Generic (PLEG): container finished" podID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerID="5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48" exitCode=0 Oct 07 21:46:02 crc kubenswrapper[4926]: I1007 21:46:02.595900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerDied","Data":"5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48"} Oct 07 21:46:03 crc kubenswrapper[4926]: I1007 21:46:03.608712 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerStarted","Data":"8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c"} Oct 07 21:46:03 crc kubenswrapper[4926]: I1007 21:46:03.637657 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-svwlj" podStartSLOduration=3.18427668 podStartE2EDuration="5.637630804s" podCreationTimestamp="2025-10-07 21:45:58 +0000 UTC" firstStartedPulling="2025-10-07 21:46:00.564465869 +0000 UTC m=+3010.603047049" lastFinishedPulling="2025-10-07 21:46:03.017820023 +0000 UTC m=+3013.056401173" observedRunningTime="2025-10-07 21:46:03.629381455 +0000 UTC m=+3013.667962625" watchObservedRunningTime="2025-10-07 21:46:03.637630804 +0000 UTC m=+3013.676211994" Oct 07 21:46:05 crc kubenswrapper[4926]: I1007 21:46:05.703433 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:46:05 crc kubenswrapper[4926]: E1007 21:46:05.711819 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:46:09 crc kubenswrapper[4926]: I1007 21:46:09.151763 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:46:09 crc kubenswrapper[4926]: I1007 21:46:09.152300 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:46:09 crc kubenswrapper[4926]: I1007 21:46:09.201850 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:46:09 crc kubenswrapper[4926]: I1007 21:46:09.732775 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:46:09 crc kubenswrapper[4926]: I1007 21:46:09.822889 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-svwlj"] Oct 07 21:46:11 crc kubenswrapper[4926]: I1007 21:46:11.681878 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-svwlj" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="registry-server" containerID="cri-o://8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c" gracePeriod=2 Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.231019 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.273624 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-utilities\") pod \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.274007 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8nll\" (UniqueName: \"kubernetes.io/projected/b72e553d-26ac-460e-94da-a5ffc59f9a8a-kube-api-access-j8nll\") pod \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.274358 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-catalog-content\") pod \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\" (UID: \"b72e553d-26ac-460e-94da-a5ffc59f9a8a\") " Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.275293 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-utilities" (OuterVolumeSpecName: "utilities") pod "b72e553d-26ac-460e-94da-a5ffc59f9a8a" (UID: "b72e553d-26ac-460e-94da-a5ffc59f9a8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.275574 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.279878 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72e553d-26ac-460e-94da-a5ffc59f9a8a-kube-api-access-j8nll" (OuterVolumeSpecName: "kube-api-access-j8nll") pod "b72e553d-26ac-460e-94da-a5ffc59f9a8a" (UID: "b72e553d-26ac-460e-94da-a5ffc59f9a8a"). InnerVolumeSpecName "kube-api-access-j8nll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.285872 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b72e553d-26ac-460e-94da-a5ffc59f9a8a" (UID: "b72e553d-26ac-460e-94da-a5ffc59f9a8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.376830 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8nll\" (UniqueName: \"kubernetes.io/projected/b72e553d-26ac-460e-94da-a5ffc59f9a8a-kube-api-access-j8nll\") on node \"crc\" DevicePath \"\"" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.376878 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e553d-26ac-460e-94da-a5ffc59f9a8a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.694187 4926 generic.go:334] "Generic (PLEG): container finished" podID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerID="8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c" exitCode=0 Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.695467 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svwlj" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.695904 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerDied","Data":"8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c"} Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.695946 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svwlj" event={"ID":"b72e553d-26ac-460e-94da-a5ffc59f9a8a","Type":"ContainerDied","Data":"48596de64612e8cddc9920884858855d7953683b4768e002bf9b773c0a17c22d"} Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.695964 4926 scope.go:117] "RemoveContainer" containerID="8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.721356 4926 scope.go:117] "RemoveContainer" containerID="5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.732746 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-svwlj"] Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.747437 4926 scope.go:117] "RemoveContainer" containerID="5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.750071 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-svwlj"] Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.789031 4926 scope.go:117] "RemoveContainer" containerID="8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c" Oct 07 21:46:12 crc kubenswrapper[4926]: E1007 21:46:12.790716 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c\": container with ID starting with 8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c not found: ID does not exist" containerID="8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.790770 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c"} err="failed to get container status \"8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c\": rpc error: code = NotFound desc = could not find container \"8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c\": container with ID starting with 8a41f2a51366da527037c4347e132d0c234de53c703182e6eeed3c5f71dab42c not found: ID does not exist" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.790853 4926 scope.go:117] "RemoveContainer" containerID="5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48" Oct 07 21:46:12 crc kubenswrapper[4926]: E1007 21:46:12.791280 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48\": container with ID starting with 5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48 not found: ID does not exist" containerID="5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.791324 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48"} err="failed to get container status \"5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48\": rpc error: code = NotFound desc = could not find container \"5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48\": container with ID starting with 5d5ea201bafef3a916378634f281638cc40b0859f3a74ad02596a0dcd906ab48 not found: ID does not exist" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.791362 4926 scope.go:117] "RemoveContainer" containerID="5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50" Oct 07 21:46:12 crc kubenswrapper[4926]: E1007 21:46:12.791707 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50\": container with ID starting with 5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50 not found: ID does not exist" containerID="5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50" Oct 07 21:46:12 crc kubenswrapper[4926]: I1007 21:46:12.791747 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50"} err="failed to get container status \"5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50\": rpc error: code = NotFound desc = could not find container \"5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50\": container with ID starting with 5aba6caf87b7c2074b05c353efcac3f7933f851f42dc87d8dd955c6a22f68d50 not found: ID does not exist" Oct 07 21:46:14 crc kubenswrapper[4926]: I1007 21:46:14.694080 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" path="/var/lib/kubelet/pods/b72e553d-26ac-460e-94da-a5ffc59f9a8a/volumes" Oct 07 21:46:18 crc kubenswrapper[4926]: I1007 21:46:18.679139 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:46:18 crc kubenswrapper[4926]: E1007 21:46:18.680502 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:46:32 crc kubenswrapper[4926]: I1007 21:46:32.679473 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:46:32 crc kubenswrapper[4926]: E1007 21:46:32.680478 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:46:43 crc kubenswrapper[4926]: I1007 21:46:43.679814 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:46:43 crc kubenswrapper[4926]: E1007 21:46:43.682320 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:46:54 crc kubenswrapper[4926]: I1007 21:46:54.680053 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:46:54 crc kubenswrapper[4926]: E1007 21:46:54.680701 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:47:06 crc kubenswrapper[4926]: I1007 21:47:06.679907 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:47:06 crc kubenswrapper[4926]: E1007 21:47:06.682154 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:47:18 crc kubenswrapper[4926]: I1007 21:47:18.680452 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:47:18 crc kubenswrapper[4926]: E1007 21:47:18.681656 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:47:32 crc kubenswrapper[4926]: I1007 21:47:32.679830 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:47:32 crc kubenswrapper[4926]: E1007 21:47:32.680588 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:47:47 crc kubenswrapper[4926]: I1007 21:47:47.678973 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:47:47 crc kubenswrapper[4926]: E1007 21:47:47.679986 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:48:02 crc kubenswrapper[4926]: I1007 21:48:02.678900 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:48:02 crc kubenswrapper[4926]: E1007 21:48:02.680000 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:48:15 crc kubenswrapper[4926]: I1007 21:48:15.679561 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:48:15 crc kubenswrapper[4926]: E1007 21:48:15.680561 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:48:30 crc kubenswrapper[4926]: I1007 21:48:30.692091 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:48:30 crc kubenswrapper[4926]: E1007 21:48:30.693082 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:48:43 crc kubenswrapper[4926]: I1007 21:48:43.679874 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:48:43 crc kubenswrapper[4926]: E1007 21:48:43.681239 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:48:54 crc kubenswrapper[4926]: I1007 21:48:54.679564 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:48:54 crc kubenswrapper[4926]: E1007 21:48:54.680630 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:49:08 crc kubenswrapper[4926]: I1007 21:49:08.679803 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:49:08 crc kubenswrapper[4926]: E1007 21:49:08.680827 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:49:20 crc kubenswrapper[4926]: I1007 21:49:20.698334 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:49:20 crc kubenswrapper[4926]: E1007 21:49:20.699610 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:49:34 crc kubenswrapper[4926]: I1007 21:49:34.679442 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:49:35 crc kubenswrapper[4926]: I1007 21:49:35.085480 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"4db052a92df06ad93ca8525c42aaed0f766db998ce6adff7f06bde2239b4ddcf"} Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.136976 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4b9t4"] Oct 07 21:50:34 crc kubenswrapper[4926]: E1007 21:50:34.137927 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="registry-server" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.137942 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="registry-server" Oct 07 21:50:34 crc kubenswrapper[4926]: E1007 21:50:34.137968 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="extract-content" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.137977 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="extract-content" Oct 07 21:50:34 crc kubenswrapper[4926]: E1007 21:50:34.138001 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="extract-utilities" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.138010 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="extract-utilities" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.138251 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72e553d-26ac-460e-94da-a5ffc59f9a8a" containerName="registry-server" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.141654 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.164693 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4b9t4"] Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.301450 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdwl6\" (UniqueName: \"kubernetes.io/projected/6a3e4134-4cfd-4129-b2da-fd9aecd39447-kube-api-access-xdwl6\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.301823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-catalog-content\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.302061 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-utilities\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.404006 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdwl6\" (UniqueName: \"kubernetes.io/projected/6a3e4134-4cfd-4129-b2da-fd9aecd39447-kube-api-access-xdwl6\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.404084 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-catalog-content\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.404175 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-utilities\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.404644 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-catalog-content\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.404678 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-utilities\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.423920 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdwl6\" (UniqueName: \"kubernetes.io/projected/6a3e4134-4cfd-4129-b2da-fd9aecd39447-kube-api-access-xdwl6\") pod \"certified-operators-4b9t4\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:34 crc kubenswrapper[4926]: I1007 21:50:34.466467 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:35 crc kubenswrapper[4926]: I1007 21:50:35.480207 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4b9t4"] Oct 07 21:50:35 crc kubenswrapper[4926]: I1007 21:50:35.781382 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerID="6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70" exitCode=0 Oct 07 21:50:35 crc kubenswrapper[4926]: I1007 21:50:35.781488 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4b9t4" event={"ID":"6a3e4134-4cfd-4129-b2da-fd9aecd39447","Type":"ContainerDied","Data":"6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70"} Oct 07 21:50:35 crc kubenswrapper[4926]: I1007 21:50:35.781759 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4b9t4" event={"ID":"6a3e4134-4cfd-4129-b2da-fd9aecd39447","Type":"ContainerStarted","Data":"80f3dc8eae0545a3a02e67a7cfb8a63ca68c4c920b7ac562f7efe371e7b23e56"} Oct 07 21:50:35 crc kubenswrapper[4926]: I1007 21:50:35.783355 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:50:37 crc kubenswrapper[4926]: I1007 21:50:37.811826 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerID="572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4" exitCode=0 Oct 07 21:50:37 crc kubenswrapper[4926]: I1007 21:50:37.812229 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4b9t4" event={"ID":"6a3e4134-4cfd-4129-b2da-fd9aecd39447","Type":"ContainerDied","Data":"572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4"} Oct 07 21:50:38 crc kubenswrapper[4926]: I1007 21:50:38.825855 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4b9t4" event={"ID":"6a3e4134-4cfd-4129-b2da-fd9aecd39447","Type":"ContainerStarted","Data":"feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109"} Oct 07 21:50:44 crc kubenswrapper[4926]: I1007 21:50:44.467706 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:44 crc kubenswrapper[4926]: I1007 21:50:44.470337 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:44 crc kubenswrapper[4926]: I1007 21:50:44.558508 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:44 crc kubenswrapper[4926]: I1007 21:50:44.593141 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4b9t4" podStartSLOduration=8.059698908 podStartE2EDuration="10.593109954s" podCreationTimestamp="2025-10-07 21:50:34 +0000 UTC" firstStartedPulling="2025-10-07 21:50:35.78290365 +0000 UTC m=+3285.821484800" lastFinishedPulling="2025-10-07 21:50:38.316314696 +0000 UTC m=+3288.354895846" observedRunningTime="2025-10-07 21:50:38.854492952 +0000 UTC m=+3288.893074112" watchObservedRunningTime="2025-10-07 21:50:44.593109954 +0000 UTC m=+3294.631691144" Oct 07 21:50:44 crc kubenswrapper[4926]: I1007 21:50:44.963227 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:45 crc kubenswrapper[4926]: I1007 21:50:45.047688 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4b9t4"] Oct 07 21:50:46 crc kubenswrapper[4926]: I1007 21:50:46.919635 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4b9t4" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="registry-server" containerID="cri-o://feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109" gracePeriod=2 Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.428240 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.588432 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-utilities\") pod \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.588989 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-catalog-content\") pod \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.589219 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdwl6\" (UniqueName: \"kubernetes.io/projected/6a3e4134-4cfd-4129-b2da-fd9aecd39447-kube-api-access-xdwl6\") pod \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\" (UID: \"6a3e4134-4cfd-4129-b2da-fd9aecd39447\") " Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.589270 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-utilities" (OuterVolumeSpecName: "utilities") pod "6a3e4134-4cfd-4129-b2da-fd9aecd39447" (UID: "6a3e4134-4cfd-4129-b2da-fd9aecd39447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.589813 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.596091 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3e4134-4cfd-4129-b2da-fd9aecd39447-kube-api-access-xdwl6" (OuterVolumeSpecName: "kube-api-access-xdwl6") pod "6a3e4134-4cfd-4129-b2da-fd9aecd39447" (UID: "6a3e4134-4cfd-4129-b2da-fd9aecd39447"). InnerVolumeSpecName "kube-api-access-xdwl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.635500 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a3e4134-4cfd-4129-b2da-fd9aecd39447" (UID: "6a3e4134-4cfd-4129-b2da-fd9aecd39447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.692255 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3e4134-4cfd-4129-b2da-fd9aecd39447-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.692323 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdwl6\" (UniqueName: \"kubernetes.io/projected/6a3e4134-4cfd-4129-b2da-fd9aecd39447-kube-api-access-xdwl6\") on node \"crc\" DevicePath \"\"" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.939802 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerID="feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109" exitCode=0 Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.939851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4b9t4" event={"ID":"6a3e4134-4cfd-4129-b2da-fd9aecd39447","Type":"ContainerDied","Data":"feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109"} Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.939879 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4b9t4" event={"ID":"6a3e4134-4cfd-4129-b2da-fd9aecd39447","Type":"ContainerDied","Data":"80f3dc8eae0545a3a02e67a7cfb8a63ca68c4c920b7ac562f7efe371e7b23e56"} Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.939895 4926 scope.go:117] "RemoveContainer" containerID="feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.940021 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4b9t4" Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.989470 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4b9t4"] Oct 07 21:50:47 crc kubenswrapper[4926]: I1007 21:50:47.992250 4926 scope.go:117] "RemoveContainer" containerID="572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.005940 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4b9t4"] Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.026823 4926 scope.go:117] "RemoveContainer" containerID="6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.070078 4926 scope.go:117] "RemoveContainer" containerID="feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109" Oct 07 21:50:48 crc kubenswrapper[4926]: E1007 21:50:48.070575 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109\": container with ID starting with feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109 not found: ID does not exist" containerID="feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.070617 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109"} err="failed to get container status \"feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109\": rpc error: code = NotFound desc = could not find container \"feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109\": container with ID starting with feb9f79febb823f5f78b84804f6f8c3168483b303e33fe7a7042d5ac9ae23109 not found: ID does not exist" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.070645 4926 scope.go:117] "RemoveContainer" containerID="572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4" Oct 07 21:50:48 crc kubenswrapper[4926]: E1007 21:50:48.071032 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4\": container with ID starting with 572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4 not found: ID does not exist" containerID="572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.071068 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4"} err="failed to get container status \"572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4\": rpc error: code = NotFound desc = could not find container \"572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4\": container with ID starting with 572f6926a33f9519e9d6da7835e66a1521833c8592a6b86fb1d0986a81008ec4 not found: ID does not exist" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.071085 4926 scope.go:117] "RemoveContainer" containerID="6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70" Oct 07 21:50:48 crc kubenswrapper[4926]: E1007 21:50:48.071345 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70\": container with ID starting with 6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70 not found: ID does not exist" containerID="6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.071379 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70"} err="failed to get container status \"6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70\": rpc error: code = NotFound desc = could not find container \"6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70\": container with ID starting with 6c246dbaaa7d57de4079cbd229df95f4e1e50e47861a831e0ab731d0c4826f70 not found: ID does not exist" Oct 07 21:50:48 crc kubenswrapper[4926]: I1007 21:50:48.691764 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" path="/var/lib/kubelet/pods/6a3e4134-4cfd-4129-b2da-fd9aecd39447/volumes" Oct 07 21:52:03 crc kubenswrapper[4926]: I1007 21:52:03.209220 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:52:03 crc kubenswrapper[4926]: I1007 21:52:03.209873 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:52:33 crc kubenswrapper[4926]: I1007 21:52:33.209349 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:52:33 crc kubenswrapper[4926]: I1007 21:52:33.209926 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.423911 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lltqd"] Oct 07 21:52:47 crc kubenswrapper[4926]: E1007 21:52:47.425012 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="extract-utilities" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.425029 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="extract-utilities" Oct 07 21:52:47 crc kubenswrapper[4926]: E1007 21:52:47.425074 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="registry-server" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.425083 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="registry-server" Oct 07 21:52:47 crc kubenswrapper[4926]: E1007 21:52:47.425111 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="extract-content" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.425120 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="extract-content" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.425403 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3e4134-4cfd-4129-b2da-fd9aecd39447" containerName="registry-server" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.427238 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.434023 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lltqd"] Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.601081 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-utilities\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.601148 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75nmx\" (UniqueName: \"kubernetes.io/projected/1679d7e1-9062-449d-8aab-2acade957067-kube-api-access-75nmx\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.601308 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-catalog-content\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.703462 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75nmx\" (UniqueName: \"kubernetes.io/projected/1679d7e1-9062-449d-8aab-2acade957067-kube-api-access-75nmx\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.703570 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-catalog-content\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.703667 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-utilities\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.704308 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-catalog-content\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.704324 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-utilities\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.723924 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75nmx\" (UniqueName: \"kubernetes.io/projected/1679d7e1-9062-449d-8aab-2acade957067-kube-api-access-75nmx\") pod \"redhat-operators-lltqd\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:47 crc kubenswrapper[4926]: I1007 21:52:47.750069 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:48 crc kubenswrapper[4926]: I1007 21:52:48.360701 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lltqd"] Oct 07 21:52:48 crc kubenswrapper[4926]: I1007 21:52:48.394381 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerStarted","Data":"fa66a700adc3c76f7b2b90fdfec083963da6c528c1ba364935a0e4fedc474a26"} Oct 07 21:52:49 crc kubenswrapper[4926]: I1007 21:52:49.407731 4926 generic.go:334] "Generic (PLEG): container finished" podID="1679d7e1-9062-449d-8aab-2acade957067" containerID="154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e" exitCode=0 Oct 07 21:52:49 crc kubenswrapper[4926]: I1007 21:52:49.407801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerDied","Data":"154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e"} Oct 07 21:52:51 crc kubenswrapper[4926]: I1007 21:52:51.428301 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerStarted","Data":"a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271"} Oct 07 21:52:52 crc kubenswrapper[4926]: I1007 21:52:52.442932 4926 generic.go:334] "Generic (PLEG): container finished" podID="1679d7e1-9062-449d-8aab-2acade957067" containerID="a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271" exitCode=0 Oct 07 21:52:52 crc kubenswrapper[4926]: I1007 21:52:52.443072 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerDied","Data":"a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271"} Oct 07 21:52:53 crc kubenswrapper[4926]: I1007 21:52:53.462141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerStarted","Data":"d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654"} Oct 07 21:52:53 crc kubenswrapper[4926]: I1007 21:52:53.489104 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lltqd" podStartSLOduration=3.027421685 podStartE2EDuration="6.489083873s" podCreationTimestamp="2025-10-07 21:52:47 +0000 UTC" firstStartedPulling="2025-10-07 21:52:49.411239879 +0000 UTC m=+3419.449821079" lastFinishedPulling="2025-10-07 21:52:52.872902117 +0000 UTC m=+3422.911483267" observedRunningTime="2025-10-07 21:52:53.484249163 +0000 UTC m=+3423.522830313" watchObservedRunningTime="2025-10-07 21:52:53.489083873 +0000 UTC m=+3423.527665023" Oct 07 21:52:57 crc kubenswrapper[4926]: I1007 21:52:57.750686 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:57 crc kubenswrapper[4926]: I1007 21:52:57.751549 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:57 crc kubenswrapper[4926]: I1007 21:52:57.814610 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:58 crc kubenswrapper[4926]: I1007 21:52:58.577824 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:52:58 crc kubenswrapper[4926]: I1007 21:52:58.628450 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lltqd"] Oct 07 21:53:00 crc kubenswrapper[4926]: I1007 21:53:00.532143 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lltqd" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="registry-server" containerID="cri-o://d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654" gracePeriod=2 Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.049784 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.208531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75nmx\" (UniqueName: \"kubernetes.io/projected/1679d7e1-9062-449d-8aab-2acade957067-kube-api-access-75nmx\") pod \"1679d7e1-9062-449d-8aab-2acade957067\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.208753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-catalog-content\") pod \"1679d7e1-9062-449d-8aab-2acade957067\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.208793 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-utilities\") pod \"1679d7e1-9062-449d-8aab-2acade957067\" (UID: \"1679d7e1-9062-449d-8aab-2acade957067\") " Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.209738 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-utilities" (OuterVolumeSpecName: "utilities") pod "1679d7e1-9062-449d-8aab-2acade957067" (UID: "1679d7e1-9062-449d-8aab-2acade957067"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.217186 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1679d7e1-9062-449d-8aab-2acade957067-kube-api-access-75nmx" (OuterVolumeSpecName: "kube-api-access-75nmx") pod "1679d7e1-9062-449d-8aab-2acade957067" (UID: "1679d7e1-9062-449d-8aab-2acade957067"). InnerVolumeSpecName "kube-api-access-75nmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.311601 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.311697 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75nmx\" (UniqueName: \"kubernetes.io/projected/1679d7e1-9062-449d-8aab-2acade957067-kube-api-access-75nmx\") on node \"crc\" DevicePath \"\"" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.329936 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1679d7e1-9062-449d-8aab-2acade957067" (UID: "1679d7e1-9062-449d-8aab-2acade957067"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.414074 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1679d7e1-9062-449d-8aab-2acade957067-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.546801 4926 generic.go:334] "Generic (PLEG): container finished" podID="1679d7e1-9062-449d-8aab-2acade957067" containerID="d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654" exitCode=0 Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.546851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerDied","Data":"d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654"} Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.546886 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lltqd" event={"ID":"1679d7e1-9062-449d-8aab-2acade957067","Type":"ContainerDied","Data":"fa66a700adc3c76f7b2b90fdfec083963da6c528c1ba364935a0e4fedc474a26"} Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.546903 4926 scope.go:117] "RemoveContainer" containerID="d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.547011 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lltqd" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.570905 4926 scope.go:117] "RemoveContainer" containerID="a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.598188 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lltqd"] Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.609772 4926 scope.go:117] "RemoveContainer" containerID="154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.610166 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lltqd"] Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.657733 4926 scope.go:117] "RemoveContainer" containerID="d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654" Oct 07 21:53:01 crc kubenswrapper[4926]: E1007 21:53:01.658275 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654\": container with ID starting with d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654 not found: ID does not exist" containerID="d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.658335 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654"} err="failed to get container status \"d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654\": rpc error: code = NotFound desc = could not find container \"d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654\": container with ID starting with d8357ce5b4a32cb07fd2e1548946f8005b495fcac4496b45d292fbd18bb89654 not found: ID does not exist" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.658373 4926 scope.go:117] "RemoveContainer" containerID="a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271" Oct 07 21:53:01 crc kubenswrapper[4926]: E1007 21:53:01.658749 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271\": container with ID starting with a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271 not found: ID does not exist" containerID="a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.658792 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271"} err="failed to get container status \"a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271\": rpc error: code = NotFound desc = could not find container \"a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271\": container with ID starting with a8711d11dc1d82b1cb58070503f3860390e37c1862da7f81f1ec32d2be147271 not found: ID does not exist" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.658825 4926 scope.go:117] "RemoveContainer" containerID="154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e" Oct 07 21:53:01 crc kubenswrapper[4926]: E1007 21:53:01.659121 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e\": container with ID starting with 154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e not found: ID does not exist" containerID="154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e" Oct 07 21:53:01 crc kubenswrapper[4926]: I1007 21:53:01.659158 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e"} err="failed to get container status \"154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e\": rpc error: code = NotFound desc = could not find container \"154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e\": container with ID starting with 154bc3368e2d5b7ddbbc8fb35f43bc4ff528b2c17f3c4dbf493671de37effd9e not found: ID does not exist" Oct 07 21:53:02 crc kubenswrapper[4926]: I1007 21:53:02.700002 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1679d7e1-9062-449d-8aab-2acade957067" path="/var/lib/kubelet/pods/1679d7e1-9062-449d-8aab-2acade957067/volumes" Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.209348 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.209413 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.209469 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.210857 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4db052a92df06ad93ca8525c42aaed0f766db998ce6adff7f06bde2239b4ddcf"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.210933 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://4db052a92df06ad93ca8525c42aaed0f766db998ce6adff7f06bde2239b4ddcf" gracePeriod=600 Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.579514 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="4db052a92df06ad93ca8525c42aaed0f766db998ce6adff7f06bde2239b4ddcf" exitCode=0 Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.579587 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"4db052a92df06ad93ca8525c42aaed0f766db998ce6adff7f06bde2239b4ddcf"} Oct 07 21:53:03 crc kubenswrapper[4926]: I1007 21:53:03.579941 4926 scope.go:117] "RemoveContainer" containerID="36d2e104477cc7564bcd506828800651d5abd18c381dfc7ebad020d2fd0d4ca2" Oct 07 21:53:04 crc kubenswrapper[4926]: I1007 21:53:04.608473 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d"} Oct 07 21:55:03 crc kubenswrapper[4926]: I1007 21:55:03.209908 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:55:03 crc kubenswrapper[4926]: I1007 21:55:03.210606 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:55:33 crc kubenswrapper[4926]: I1007 21:55:33.209789 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:55:33 crc kubenswrapper[4926]: I1007 21:55:33.210476 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.210288 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.211067 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.211156 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.212101 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.212183 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" gracePeriod=600 Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.765058 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" exitCode=0 Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.765103 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d"} Oct 07 21:56:03 crc kubenswrapper[4926]: I1007 21:56:03.765147 4926 scope.go:117] "RemoveContainer" containerID="4db052a92df06ad93ca8525c42aaed0f766db998ce6adff7f06bde2239b4ddcf" Oct 07 21:56:03 crc kubenswrapper[4926]: E1007 21:56:03.899805 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:56:04 crc kubenswrapper[4926]: I1007 21:56:04.783785 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:56:04 crc kubenswrapper[4926]: E1007 21:56:04.784297 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:56:15 crc kubenswrapper[4926]: I1007 21:56:15.680137 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:56:15 crc kubenswrapper[4926]: E1007 21:56:15.681676 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.610644 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2x"] Oct 07 21:56:23 crc kubenswrapper[4926]: E1007 21:56:23.612373 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="registry-server" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.612397 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="registry-server" Oct 07 21:56:23 crc kubenswrapper[4926]: E1007 21:56:23.612430 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="extract-utilities" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.612445 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="extract-utilities" Oct 07 21:56:23 crc kubenswrapper[4926]: E1007 21:56:23.612511 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="extract-content" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.612526 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="extract-content" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.612952 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1679d7e1-9062-449d-8aab-2acade957067" containerName="registry-server" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.616300 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.627166 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2x"] Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.719137 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-catalog-content\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.719441 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94pwf\" (UniqueName: \"kubernetes.io/projected/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-kube-api-access-94pwf\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.719561 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-utilities\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.821135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94pwf\" (UniqueName: \"kubernetes.io/projected/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-kube-api-access-94pwf\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.821646 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-utilities\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.821809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-catalog-content\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.822911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-catalog-content\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.823022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-utilities\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.855618 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94pwf\" (UniqueName: \"kubernetes.io/projected/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-kube-api-access-94pwf\") pod \"redhat-marketplace-7cz2x\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:23 crc kubenswrapper[4926]: I1007 21:56:23.968555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:24 crc kubenswrapper[4926]: I1007 21:56:24.413587 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2x"] Oct 07 21:56:25 crc kubenswrapper[4926]: I1007 21:56:25.004408 4926 generic.go:334] "Generic (PLEG): container finished" podID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerID="786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3" exitCode=0 Oct 07 21:56:25 crc kubenswrapper[4926]: I1007 21:56:25.004494 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerDied","Data":"786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3"} Oct 07 21:56:25 crc kubenswrapper[4926]: I1007 21:56:25.004748 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerStarted","Data":"7195fc29f088e1a5ba62c088e56a30844215223df797536b9d52a41278e4e08f"} Oct 07 21:56:25 crc kubenswrapper[4926]: I1007 21:56:25.006802 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 21:56:26 crc kubenswrapper[4926]: I1007 21:56:26.018467 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerStarted","Data":"1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2"} Oct 07 21:56:27 crc kubenswrapper[4926]: I1007 21:56:27.032950 4926 generic.go:334] "Generic (PLEG): container finished" podID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerID="1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2" exitCode=0 Oct 07 21:56:27 crc kubenswrapper[4926]: I1007 21:56:27.033017 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerDied","Data":"1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2"} Oct 07 21:56:28 crc kubenswrapper[4926]: I1007 21:56:28.048620 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerStarted","Data":"4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403"} Oct 07 21:56:28 crc kubenswrapper[4926]: I1007 21:56:28.078058 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7cz2x" podStartSLOduration=2.624286488 podStartE2EDuration="5.078032599s" podCreationTimestamp="2025-10-07 21:56:23 +0000 UTC" firstStartedPulling="2025-10-07 21:56:25.006570664 +0000 UTC m=+3635.045151814" lastFinishedPulling="2025-10-07 21:56:27.460316755 +0000 UTC m=+3637.498897925" observedRunningTime="2025-10-07 21:56:28.070392548 +0000 UTC m=+3638.108973718" watchObservedRunningTime="2025-10-07 21:56:28.078032599 +0000 UTC m=+3638.116613769" Oct 07 21:56:29 crc kubenswrapper[4926]: I1007 21:56:29.679971 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:56:29 crc kubenswrapper[4926]: E1007 21:56:29.684851 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:56:33 crc kubenswrapper[4926]: I1007 21:56:33.968802 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:33 crc kubenswrapper[4926]: I1007 21:56:33.970236 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:34 crc kubenswrapper[4926]: I1007 21:56:34.029483 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:34 crc kubenswrapper[4926]: I1007 21:56:34.174431 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:34 crc kubenswrapper[4926]: I1007 21:56:34.273546 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2x"] Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.139181 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7cz2x" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="registry-server" containerID="cri-o://4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403" gracePeriod=2 Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.636539 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.711927 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-catalog-content\") pod \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.712122 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94pwf\" (UniqueName: \"kubernetes.io/projected/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-kube-api-access-94pwf\") pod \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.712158 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-utilities\") pod \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\" (UID: \"2dbd6118-c3f3-4e6f-9a88-9c98462b5343\") " Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.712980 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-utilities" (OuterVolumeSpecName: "utilities") pod "2dbd6118-c3f3-4e6f-9a88-9c98462b5343" (UID: "2dbd6118-c3f3-4e6f-9a88-9c98462b5343"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.719282 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-kube-api-access-94pwf" (OuterVolumeSpecName: "kube-api-access-94pwf") pod "2dbd6118-c3f3-4e6f-9a88-9c98462b5343" (UID: "2dbd6118-c3f3-4e6f-9a88-9c98462b5343"). InnerVolumeSpecName "kube-api-access-94pwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.728602 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dbd6118-c3f3-4e6f-9a88-9c98462b5343" (UID: "2dbd6118-c3f3-4e6f-9a88-9c98462b5343"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.816176 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.816243 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94pwf\" (UniqueName: \"kubernetes.io/projected/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-kube-api-access-94pwf\") on node \"crc\" DevicePath \"\"" Oct 07 21:56:36 crc kubenswrapper[4926]: I1007 21:56:36.816263 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbd6118-c3f3-4e6f-9a88-9c98462b5343-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.151960 4926 generic.go:334] "Generic (PLEG): container finished" podID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerID="4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403" exitCode=0 Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.152003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerDied","Data":"4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403"} Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.152036 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2x" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.152059 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2x" event={"ID":"2dbd6118-c3f3-4e6f-9a88-9c98462b5343","Type":"ContainerDied","Data":"7195fc29f088e1a5ba62c088e56a30844215223df797536b9d52a41278e4e08f"} Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.152111 4926 scope.go:117] "RemoveContainer" containerID="4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.176352 4926 scope.go:117] "RemoveContainer" containerID="1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.190119 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2x"] Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.197758 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2x"] Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.220148 4926 scope.go:117] "RemoveContainer" containerID="786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.270560 4926 scope.go:117] "RemoveContainer" containerID="4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403" Oct 07 21:56:37 crc kubenswrapper[4926]: E1007 21:56:37.271012 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403\": container with ID starting with 4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403 not found: ID does not exist" containerID="4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.271077 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403"} err="failed to get container status \"4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403\": rpc error: code = NotFound desc = could not find container \"4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403\": container with ID starting with 4c031ea386a2760d69fc5f066dd032b895df3c633817051ea083e2c714dd9403 not found: ID does not exist" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.271120 4926 scope.go:117] "RemoveContainer" containerID="1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2" Oct 07 21:56:37 crc kubenswrapper[4926]: E1007 21:56:37.271643 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2\": container with ID starting with 1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2 not found: ID does not exist" containerID="1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.271681 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2"} err="failed to get container status \"1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2\": rpc error: code = NotFound desc = could not find container \"1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2\": container with ID starting with 1b3074c218d28c519cb66ebd576ce7898497b7921e77a19a3ebb45cd2f4889e2 not found: ID does not exist" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.271707 4926 scope.go:117] "RemoveContainer" containerID="786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3" Oct 07 21:56:37 crc kubenswrapper[4926]: E1007 21:56:37.271994 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3\": container with ID starting with 786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3 not found: ID does not exist" containerID="786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3" Oct 07 21:56:37 crc kubenswrapper[4926]: I1007 21:56:37.272025 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3"} err="failed to get container status \"786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3\": rpc error: code = NotFound desc = could not find container \"786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3\": container with ID starting with 786dee7dff08bed6336e2cb31acaccd383745d030c2cf854b7b2019c96be53f3 not found: ID does not exist" Oct 07 21:56:38 crc kubenswrapper[4926]: I1007 21:56:38.697430 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" path="/var/lib/kubelet/pods/2dbd6118-c3f3-4e6f-9a88-9c98462b5343/volumes" Oct 07 21:56:44 crc kubenswrapper[4926]: I1007 21:56:44.679835 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:56:44 crc kubenswrapper[4926]: E1007 21:56:44.680660 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:56:55 crc kubenswrapper[4926]: I1007 21:56:55.679053 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:56:55 crc kubenswrapper[4926]: E1007 21:56:55.680153 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:57:10 crc kubenswrapper[4926]: I1007 21:57:10.695101 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:57:10 crc kubenswrapper[4926]: E1007 21:57:10.698987 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:57:24 crc kubenswrapper[4926]: I1007 21:57:24.680246 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:57:24 crc kubenswrapper[4926]: E1007 21:57:24.681521 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:57:36 crc kubenswrapper[4926]: I1007 21:57:36.681058 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:57:36 crc kubenswrapper[4926]: E1007 21:57:36.682100 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:57:50 crc kubenswrapper[4926]: I1007 21:57:50.728062 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:57:50 crc kubenswrapper[4926]: E1007 21:57:50.728914 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:58:01 crc kubenswrapper[4926]: I1007 21:58:01.679068 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:58:01 crc kubenswrapper[4926]: E1007 21:58:01.680014 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:58:12 crc kubenswrapper[4926]: I1007 21:58:12.678583 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:58:12 crc kubenswrapper[4926]: E1007 21:58:12.679397 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:58:17 crc kubenswrapper[4926]: E1007 21:58:17.977506 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.58:55638->38.102.83.58:42335: write tcp 38.102.83.58:55638->38.102.83.58:42335: write: connection reset by peer Oct 07 21:58:26 crc kubenswrapper[4926]: I1007 21:58:26.679038 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:58:26 crc kubenswrapper[4926]: E1007 21:58:26.680474 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:58:41 crc kubenswrapper[4926]: I1007 21:58:41.679780 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:58:41 crc kubenswrapper[4926]: E1007 21:58:41.680767 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:58:53 crc kubenswrapper[4926]: I1007 21:58:53.680106 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:58:53 crc kubenswrapper[4926]: E1007 21:58:53.681465 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:59:06 crc kubenswrapper[4926]: I1007 21:59:06.679747 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:59:06 crc kubenswrapper[4926]: E1007 21:59:06.680654 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:59:21 crc kubenswrapper[4926]: I1007 21:59:21.679777 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:59:21 crc kubenswrapper[4926]: E1007 21:59:21.680689 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:59:36 crc kubenswrapper[4926]: I1007 21:59:36.678466 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:59:36 crc kubenswrapper[4926]: E1007 21:59:36.679342 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:59:50 crc kubenswrapper[4926]: I1007 21:59:50.695248 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 21:59:50 crc kubenswrapper[4926]: E1007 21:59:50.696639 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.327391 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kqx28"] Oct 07 21:59:59 crc kubenswrapper[4926]: E1007 21:59:59.328515 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="extract-content" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.328531 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="extract-content" Oct 07 21:59:59 crc kubenswrapper[4926]: E1007 21:59:59.328552 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="registry-server" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.328560 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="registry-server" Oct 07 21:59:59 crc kubenswrapper[4926]: E1007 21:59:59.328589 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="extract-utilities" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.328598 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="extract-utilities" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.328835 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbd6118-c3f3-4e6f-9a88-9c98462b5343" containerName="registry-server" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.330842 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.339535 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kqx28"] Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.407593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c896x\" (UniqueName: \"kubernetes.io/projected/7ed21898-af06-4ed6-81f5-4753cdcdb575-kube-api-access-c896x\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.407998 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-utilities\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.408086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-catalog-content\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.510424 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-utilities\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.510487 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-catalog-content\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.510520 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c896x\" (UniqueName: \"kubernetes.io/projected/7ed21898-af06-4ed6-81f5-4753cdcdb575-kube-api-access-c896x\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.511085 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-catalog-content\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.511305 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-utilities\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.529638 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c896x\" (UniqueName: \"kubernetes.io/projected/7ed21898-af06-4ed6-81f5-4753cdcdb575-kube-api-access-c896x\") pod \"community-operators-kqx28\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " pod="openshift-marketplace/community-operators-kqx28" Oct 07 21:59:59 crc kubenswrapper[4926]: I1007 21:59:59.649873 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.152827 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw"] Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.157353 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.159659 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.159763 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.170989 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw"] Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.210056 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kqx28"] Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.228640 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-secret-volume\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.228861 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br9wp\" (UniqueName: \"kubernetes.io/projected/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-kube-api-access-br9wp\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.228970 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-config-volume\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.331016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br9wp\" (UniqueName: \"kubernetes.io/projected/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-kube-api-access-br9wp\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.331087 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-config-volume\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.331203 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-secret-volume\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.334576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-config-volume\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.339136 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-secret-volume\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.354058 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br9wp\" (UniqueName: \"kubernetes.io/projected/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-kube-api-access-br9wp\") pod \"collect-profiles-29331240-dvddw\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.500867 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.571227 4926 generic.go:334] "Generic (PLEG): container finished" podID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerID="78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969" exitCode=0 Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.571408 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerDied","Data":"78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969"} Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.571569 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerStarted","Data":"165075c8862cf02f40cce2beb9b5cdeee181a7cd58c3b6d6775e861a82736053"} Oct 07 22:00:00 crc kubenswrapper[4926]: I1007 22:00:00.988675 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw"] Oct 07 22:00:00 crc kubenswrapper[4926]: W1007 22:00:00.995921 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbecc40b5_2aa2_4ba9_b067_cc7367ed2b78.slice/crio-65af5be481ae0c1eb03161925c176c4a248e977aca3dc72edbbd1cb58de4e9d2 WatchSource:0}: Error finding container 65af5be481ae0c1eb03161925c176c4a248e977aca3dc72edbbd1cb58de4e9d2: Status 404 returned error can't find the container with id 65af5be481ae0c1eb03161925c176c4a248e977aca3dc72edbbd1cb58de4e9d2 Oct 07 22:00:01 crc kubenswrapper[4926]: I1007 22:00:01.583185 4926 generic.go:334] "Generic (PLEG): container finished" podID="becc40b5-2aa2-4ba9-b067-cc7367ed2b78" containerID="a7110de1101b32d81a64b738b519703eec69745c9ab81416dffd9875299b990e" exitCode=0 Oct 07 22:00:01 crc kubenswrapper[4926]: I1007 22:00:01.583418 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" event={"ID":"becc40b5-2aa2-4ba9-b067-cc7367ed2b78","Type":"ContainerDied","Data":"a7110de1101b32d81a64b738b519703eec69745c9ab81416dffd9875299b990e"} Oct 07 22:00:01 crc kubenswrapper[4926]: I1007 22:00:01.583514 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" event={"ID":"becc40b5-2aa2-4ba9-b067-cc7367ed2b78","Type":"ContainerStarted","Data":"65af5be481ae0c1eb03161925c176c4a248e977aca3dc72edbbd1cb58de4e9d2"} Oct 07 22:00:01 crc kubenswrapper[4926]: I1007 22:00:01.585567 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerStarted","Data":"2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9"} Oct 07 22:00:02 crc kubenswrapper[4926]: I1007 22:00:02.596321 4926 generic.go:334] "Generic (PLEG): container finished" podID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerID="2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9" exitCode=0 Oct 07 22:00:02 crc kubenswrapper[4926]: I1007 22:00:02.596419 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerDied","Data":"2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9"} Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.007819 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.190909 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-config-volume\") pod \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.191016 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-secret-volume\") pod \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.191092 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br9wp\" (UniqueName: \"kubernetes.io/projected/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-kube-api-access-br9wp\") pod \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\" (UID: \"becc40b5-2aa2-4ba9-b067-cc7367ed2b78\") " Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.192634 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-config-volume" (OuterVolumeSpecName: "config-volume") pod "becc40b5-2aa2-4ba9-b067-cc7367ed2b78" (UID: "becc40b5-2aa2-4ba9-b067-cc7367ed2b78"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.198121 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-kube-api-access-br9wp" (OuterVolumeSpecName: "kube-api-access-br9wp") pod "becc40b5-2aa2-4ba9-b067-cc7367ed2b78" (UID: "becc40b5-2aa2-4ba9-b067-cc7367ed2b78"). InnerVolumeSpecName "kube-api-access-br9wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.206929 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "becc40b5-2aa2-4ba9-b067-cc7367ed2b78" (UID: "becc40b5-2aa2-4ba9-b067-cc7367ed2b78"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.293645 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.293848 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br9wp\" (UniqueName: \"kubernetes.io/projected/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-kube-api-access-br9wp\") on node \"crc\" DevicePath \"\"" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.293902 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/becc40b5-2aa2-4ba9-b067-cc7367ed2b78-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.608976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" event={"ID":"becc40b5-2aa2-4ba9-b067-cc7367ed2b78","Type":"ContainerDied","Data":"65af5be481ae0c1eb03161925c176c4a248e977aca3dc72edbbd1cb58de4e9d2"} Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.610227 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65af5be481ae0c1eb03161925c176c4a248e977aca3dc72edbbd1cb58de4e9d2" Oct 07 22:00:03 crc kubenswrapper[4926]: I1007 22:00:03.609228 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-dvddw" Oct 07 22:00:04 crc kubenswrapper[4926]: I1007 22:00:04.100114 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l"] Oct 07 22:00:04 crc kubenswrapper[4926]: I1007 22:00:04.107749 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331195-vw22l"] Oct 07 22:00:04 crc kubenswrapper[4926]: I1007 22:00:04.625953 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerStarted","Data":"4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404"} Oct 07 22:00:04 crc kubenswrapper[4926]: I1007 22:00:04.662187 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kqx28" podStartSLOduration=2.800284323 podStartE2EDuration="5.662168433s" podCreationTimestamp="2025-10-07 21:59:59 +0000 UTC" firstStartedPulling="2025-10-07 22:00:00.574963307 +0000 UTC m=+3850.613544457" lastFinishedPulling="2025-10-07 22:00:03.436847407 +0000 UTC m=+3853.475428567" observedRunningTime="2025-10-07 22:00:04.650567887 +0000 UTC m=+3854.689149057" watchObservedRunningTime="2025-10-07 22:00:04.662168433 +0000 UTC m=+3854.700749583" Oct 07 22:00:04 crc kubenswrapper[4926]: I1007 22:00:04.679636 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:00:04 crc kubenswrapper[4926]: E1007 22:00:04.679911 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:00:04 crc kubenswrapper[4926]: I1007 22:00:04.717697 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d1e6970-d33f-4582-9485-2a9027cbb1d9" path="/var/lib/kubelet/pods/1d1e6970-d33f-4582-9485-2a9027cbb1d9/volumes" Oct 07 22:00:05 crc kubenswrapper[4926]: I1007 22:00:05.896367 4926 scope.go:117] "RemoveContainer" containerID="3a3fd12ee6604659fb4eaa418630e6d708a54bd07d2a53fe165f8d96fe374542" Oct 07 22:00:09 crc kubenswrapper[4926]: I1007 22:00:09.650543 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:09 crc kubenswrapper[4926]: I1007 22:00:09.651304 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:09 crc kubenswrapper[4926]: I1007 22:00:09.738679 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:09 crc kubenswrapper[4926]: I1007 22:00:09.816798 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:10 crc kubenswrapper[4926]: I1007 22:00:10.325090 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kqx28"] Oct 07 22:00:11 crc kubenswrapper[4926]: I1007 22:00:11.703419 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kqx28" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="registry-server" containerID="cri-o://4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404" gracePeriod=2 Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.214334 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.304782 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c896x\" (UniqueName: \"kubernetes.io/projected/7ed21898-af06-4ed6-81f5-4753cdcdb575-kube-api-access-c896x\") pod \"7ed21898-af06-4ed6-81f5-4753cdcdb575\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.304866 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-utilities\") pod \"7ed21898-af06-4ed6-81f5-4753cdcdb575\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.304966 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-catalog-content\") pod \"7ed21898-af06-4ed6-81f5-4753cdcdb575\" (UID: \"7ed21898-af06-4ed6-81f5-4753cdcdb575\") " Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.306098 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-utilities" (OuterVolumeSpecName: "utilities") pod "7ed21898-af06-4ed6-81f5-4753cdcdb575" (UID: "7ed21898-af06-4ed6-81f5-4753cdcdb575"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.311933 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed21898-af06-4ed6-81f5-4753cdcdb575-kube-api-access-c896x" (OuterVolumeSpecName: "kube-api-access-c896x") pod "7ed21898-af06-4ed6-81f5-4753cdcdb575" (UID: "7ed21898-af06-4ed6-81f5-4753cdcdb575"). InnerVolumeSpecName "kube-api-access-c896x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.371324 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ed21898-af06-4ed6-81f5-4753cdcdb575" (UID: "7ed21898-af06-4ed6-81f5-4753cdcdb575"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.408004 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c896x\" (UniqueName: \"kubernetes.io/projected/7ed21898-af06-4ed6-81f5-4753cdcdb575-kube-api-access-c896x\") on node \"crc\" DevicePath \"\"" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.408048 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.408062 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ed21898-af06-4ed6-81f5-4753cdcdb575-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.715540 4926 generic.go:334] "Generic (PLEG): container finished" podID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerID="4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404" exitCode=0 Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.715590 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerDied","Data":"4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404"} Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.715623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqx28" event={"ID":"7ed21898-af06-4ed6-81f5-4753cdcdb575","Type":"ContainerDied","Data":"165075c8862cf02f40cce2beb9b5cdeee181a7cd58c3b6d6775e861a82736053"} Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.715646 4926 scope.go:117] "RemoveContainer" containerID="4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.717383 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqx28" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.743939 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kqx28"] Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.747301 4926 scope.go:117] "RemoveContainer" containerID="2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.756724 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kqx28"] Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.778887 4926 scope.go:117] "RemoveContainer" containerID="78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.840304 4926 scope.go:117] "RemoveContainer" containerID="4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404" Oct 07 22:00:12 crc kubenswrapper[4926]: E1007 22:00:12.841228 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404\": container with ID starting with 4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404 not found: ID does not exist" containerID="4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.841302 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404"} err="failed to get container status \"4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404\": rpc error: code = NotFound desc = could not find container \"4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404\": container with ID starting with 4103f61d6afbe3da7085458ac425969f870f97407f489c39cdeccc521821d404 not found: ID does not exist" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.841346 4926 scope.go:117] "RemoveContainer" containerID="2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9" Oct 07 22:00:12 crc kubenswrapper[4926]: E1007 22:00:12.841776 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9\": container with ID starting with 2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9 not found: ID does not exist" containerID="2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.841808 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9"} err="failed to get container status \"2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9\": rpc error: code = NotFound desc = could not find container \"2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9\": container with ID starting with 2ef45abec07b3440251c48bbb752450e9bb126185d0aea880fa258269361c8c9 not found: ID does not exist" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.841826 4926 scope.go:117] "RemoveContainer" containerID="78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969" Oct 07 22:00:12 crc kubenswrapper[4926]: E1007 22:00:12.842302 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969\": container with ID starting with 78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969 not found: ID does not exist" containerID="78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969" Oct 07 22:00:12 crc kubenswrapper[4926]: I1007 22:00:12.842341 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969"} err="failed to get container status \"78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969\": rpc error: code = NotFound desc = could not find container \"78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969\": container with ID starting with 78e57170dca35e07713610ccdcfa92240219d03f69579fb3cbd03073e566a969 not found: ID does not exist" Oct 07 22:00:14 crc kubenswrapper[4926]: I1007 22:00:14.700978 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" path="/var/lib/kubelet/pods/7ed21898-af06-4ed6-81f5-4753cdcdb575/volumes" Oct 07 22:00:18 crc kubenswrapper[4926]: I1007 22:00:18.680376 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:00:18 crc kubenswrapper[4926]: E1007 22:00:18.681320 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:00:31 crc kubenswrapper[4926]: I1007 22:00:31.679455 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:00:31 crc kubenswrapper[4926]: E1007 22:00:31.680488 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:00:46 crc kubenswrapper[4926]: I1007 22:00:46.679269 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:00:46 crc kubenswrapper[4926]: E1007 22:00:46.680372 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:00:58 crc kubenswrapper[4926]: I1007 22:00:58.683742 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:00:58 crc kubenswrapper[4926]: E1007 22:00:58.684998 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.157477 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29331241-qg8pm"] Oct 07 22:01:00 crc kubenswrapper[4926]: E1007 22:01:00.158578 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="extract-content" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.158598 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="extract-content" Oct 07 22:01:00 crc kubenswrapper[4926]: E1007 22:01:00.158629 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="extract-utilities" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.158638 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="extract-utilities" Oct 07 22:01:00 crc kubenswrapper[4926]: E1007 22:01:00.158646 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="registry-server" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.158654 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="registry-server" Oct 07 22:01:00 crc kubenswrapper[4926]: E1007 22:01:00.158699 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becc40b5-2aa2-4ba9-b067-cc7367ed2b78" containerName="collect-profiles" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.158708 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="becc40b5-2aa2-4ba9-b067-cc7367ed2b78" containerName="collect-profiles" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.158940 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="becc40b5-2aa2-4ba9-b067-cc7367ed2b78" containerName="collect-profiles" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.158973 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed21898-af06-4ed6-81f5-4753cdcdb575" containerName="registry-server" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.159884 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.178540 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331241-qg8pm"] Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.239301 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-fernet-keys\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.239355 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-combined-ca-bundle\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.239600 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc8nx\" (UniqueName: \"kubernetes.io/projected/ab4d73e0-30e0-4388-b7dc-28c97399930d-kube-api-access-hc8nx\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.240000 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-config-data\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.342262 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-fernet-keys\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.342339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-combined-ca-bundle\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.343546 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc8nx\" (UniqueName: \"kubernetes.io/projected/ab4d73e0-30e0-4388-b7dc-28c97399930d-kube-api-access-hc8nx\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.343743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-config-data\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.349991 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-combined-ca-bundle\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.350682 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-config-data\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.357177 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-fernet-keys\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.365946 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc8nx\" (UniqueName: \"kubernetes.io/projected/ab4d73e0-30e0-4388-b7dc-28c97399930d-kube-api-access-hc8nx\") pod \"keystone-cron-29331241-qg8pm\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:00 crc kubenswrapper[4926]: I1007 22:01:00.497941 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.029380 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331241-qg8pm"] Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.299767 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331241-qg8pm" event={"ID":"ab4d73e0-30e0-4388-b7dc-28c97399930d","Type":"ContainerStarted","Data":"26bc108faa0cd69c24a3d44cad235a66e89ace65aed46f737041ee3ed0ce4848"} Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.300910 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331241-qg8pm" event={"ID":"ab4d73e0-30e0-4388-b7dc-28c97399930d","Type":"ContainerStarted","Data":"57443a7fe4a4814232fcdcfc6235d567bf4c0720ba5795101ad7a99b7a37498d"} Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.332598 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29331241-qg8pm" podStartSLOduration=1.332577549 podStartE2EDuration="1.332577549s" podCreationTimestamp="2025-10-07 22:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:01:01.325491044 +0000 UTC m=+3911.364072194" watchObservedRunningTime="2025-10-07 22:01:01.332577549 +0000 UTC m=+3911.371158689" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.764909 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vhxfc"] Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.769060 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.791454 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vhxfc"] Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.876149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rzcf\" (UniqueName: \"kubernetes.io/projected/c622d8cc-2926-4e85-979b-37661d53380f-kube-api-access-6rzcf\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.876429 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-catalog-content\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.876615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-utilities\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.977996 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-catalog-content\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.978408 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-utilities\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.978578 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rzcf\" (UniqueName: \"kubernetes.io/projected/c622d8cc-2926-4e85-979b-37661d53380f-kube-api-access-6rzcf\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.978607 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-catalog-content\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:01 crc kubenswrapper[4926]: I1007 22:01:01.978932 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-utilities\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:02 crc kubenswrapper[4926]: I1007 22:01:02.005298 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rzcf\" (UniqueName: \"kubernetes.io/projected/c622d8cc-2926-4e85-979b-37661d53380f-kube-api-access-6rzcf\") pod \"certified-operators-vhxfc\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:02 crc kubenswrapper[4926]: I1007 22:01:02.095693 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:02 crc kubenswrapper[4926]: I1007 22:01:02.644148 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vhxfc"] Oct 07 22:01:02 crc kubenswrapper[4926]: W1007 22:01:02.654421 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc622d8cc_2926_4e85_979b_37661d53380f.slice/crio-54d00fc1e4f2578b09bd6e6ff1379c1b1edf2a7649b6deb5319dd6b9ddae5add WatchSource:0}: Error finding container 54d00fc1e4f2578b09bd6e6ff1379c1b1edf2a7649b6deb5319dd6b9ddae5add: Status 404 returned error can't find the container with id 54d00fc1e4f2578b09bd6e6ff1379c1b1edf2a7649b6deb5319dd6b9ddae5add Oct 07 22:01:03 crc kubenswrapper[4926]: I1007 22:01:03.320913 4926 generic.go:334] "Generic (PLEG): container finished" podID="c622d8cc-2926-4e85-979b-37661d53380f" containerID="6e79638c52e7b6ca402f6517a9cf9f3e997bb7d7b59f3822a42707850be77279" exitCode=0 Oct 07 22:01:03 crc kubenswrapper[4926]: I1007 22:01:03.321017 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerDied","Data":"6e79638c52e7b6ca402f6517a9cf9f3e997bb7d7b59f3822a42707850be77279"} Oct 07 22:01:03 crc kubenswrapper[4926]: I1007 22:01:03.321274 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerStarted","Data":"54d00fc1e4f2578b09bd6e6ff1379c1b1edf2a7649b6deb5319dd6b9ddae5add"} Oct 07 22:01:04 crc kubenswrapper[4926]: I1007 22:01:04.334988 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerStarted","Data":"891766edd3654cbd3e927447bd9d5deebbee4f9ec06526d03112d1bf133361ad"} Oct 07 22:01:05 crc kubenswrapper[4926]: I1007 22:01:05.351311 4926 generic.go:334] "Generic (PLEG): container finished" podID="ab4d73e0-30e0-4388-b7dc-28c97399930d" containerID="26bc108faa0cd69c24a3d44cad235a66e89ace65aed46f737041ee3ed0ce4848" exitCode=0 Oct 07 22:01:05 crc kubenswrapper[4926]: I1007 22:01:05.351540 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331241-qg8pm" event={"ID":"ab4d73e0-30e0-4388-b7dc-28c97399930d","Type":"ContainerDied","Data":"26bc108faa0cd69c24a3d44cad235a66e89ace65aed46f737041ee3ed0ce4848"} Oct 07 22:01:05 crc kubenswrapper[4926]: I1007 22:01:05.356134 4926 generic.go:334] "Generic (PLEG): container finished" podID="c622d8cc-2926-4e85-979b-37661d53380f" containerID="891766edd3654cbd3e927447bd9d5deebbee4f9ec06526d03112d1bf133361ad" exitCode=0 Oct 07 22:01:05 crc kubenswrapper[4926]: I1007 22:01:05.356192 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerDied","Data":"891766edd3654cbd3e927447bd9d5deebbee4f9ec06526d03112d1bf133361ad"} Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.378501 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerStarted","Data":"d75f39ba227a525d2cbfe3cc523665a988b76c030a7a0d79e632d15ce0914517"} Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.796622 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.818216 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vhxfc" podStartSLOduration=3.335063848 podStartE2EDuration="5.818181868s" podCreationTimestamp="2025-10-07 22:01:01 +0000 UTC" firstStartedPulling="2025-10-07 22:01:03.323406335 +0000 UTC m=+3913.361987525" lastFinishedPulling="2025-10-07 22:01:05.806524395 +0000 UTC m=+3915.845105545" observedRunningTime="2025-10-07 22:01:06.399581706 +0000 UTC m=+3916.438162866" watchObservedRunningTime="2025-10-07 22:01:06.818181868 +0000 UTC m=+3916.856763018" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.884087 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-config-data\") pod \"ab4d73e0-30e0-4388-b7dc-28c97399930d\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.884306 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-combined-ca-bundle\") pod \"ab4d73e0-30e0-4388-b7dc-28c97399930d\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.884343 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc8nx\" (UniqueName: \"kubernetes.io/projected/ab4d73e0-30e0-4388-b7dc-28c97399930d-kube-api-access-hc8nx\") pod \"ab4d73e0-30e0-4388-b7dc-28c97399930d\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.884429 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-fernet-keys\") pod \"ab4d73e0-30e0-4388-b7dc-28c97399930d\" (UID: \"ab4d73e0-30e0-4388-b7dc-28c97399930d\") " Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.889791 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4d73e0-30e0-4388-b7dc-28c97399930d-kube-api-access-hc8nx" (OuterVolumeSpecName: "kube-api-access-hc8nx") pod "ab4d73e0-30e0-4388-b7dc-28c97399930d" (UID: "ab4d73e0-30e0-4388-b7dc-28c97399930d"). InnerVolumeSpecName "kube-api-access-hc8nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.890260 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ab4d73e0-30e0-4388-b7dc-28c97399930d" (UID: "ab4d73e0-30e0-4388-b7dc-28c97399930d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.914180 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab4d73e0-30e0-4388-b7dc-28c97399930d" (UID: "ab4d73e0-30e0-4388-b7dc-28c97399930d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.937966 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-config-data" (OuterVolumeSpecName: "config-data") pod "ab4d73e0-30e0-4388-b7dc-28c97399930d" (UID: "ab4d73e0-30e0-4388-b7dc-28c97399930d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.986705 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.986753 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.986772 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc8nx\" (UniqueName: \"kubernetes.io/projected/ab4d73e0-30e0-4388-b7dc-28c97399930d-kube-api-access-hc8nx\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:06 crc kubenswrapper[4926]: I1007 22:01:06.986785 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab4d73e0-30e0-4388-b7dc-28c97399930d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:07 crc kubenswrapper[4926]: I1007 22:01:07.387564 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331241-qg8pm" event={"ID":"ab4d73e0-30e0-4388-b7dc-28c97399930d","Type":"ContainerDied","Data":"57443a7fe4a4814232fcdcfc6235d567bf4c0720ba5795101ad7a99b7a37498d"} Oct 07 22:01:07 crc kubenswrapper[4926]: I1007 22:01:07.387592 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331241-qg8pm" Oct 07 22:01:07 crc kubenswrapper[4926]: I1007 22:01:07.387602 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57443a7fe4a4814232fcdcfc6235d567bf4c0720ba5795101ad7a99b7a37498d" Oct 07 22:01:12 crc kubenswrapper[4926]: I1007 22:01:12.096267 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:12 crc kubenswrapper[4926]: I1007 22:01:12.096993 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:12 crc kubenswrapper[4926]: I1007 22:01:12.181356 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:12 crc kubenswrapper[4926]: I1007 22:01:12.518089 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:12 crc kubenswrapper[4926]: I1007 22:01:12.582009 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vhxfc"] Oct 07 22:01:12 crc kubenswrapper[4926]: I1007 22:01:12.679313 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:01:13 crc kubenswrapper[4926]: I1007 22:01:13.458537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"235edc795a0778f74569a034f6238258e3f846e8c83b30c0465bdc68d4b2383e"} Oct 07 22:01:14 crc kubenswrapper[4926]: I1007 22:01:14.471065 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vhxfc" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="registry-server" containerID="cri-o://d75f39ba227a525d2cbfe3cc523665a988b76c030a7a0d79e632d15ce0914517" gracePeriod=2 Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.498591 4926 generic.go:334] "Generic (PLEG): container finished" podID="c622d8cc-2926-4e85-979b-37661d53380f" containerID="d75f39ba227a525d2cbfe3cc523665a988b76c030a7a0d79e632d15ce0914517" exitCode=0 Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.498671 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerDied","Data":"d75f39ba227a525d2cbfe3cc523665a988b76c030a7a0d79e632d15ce0914517"} Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.630843 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.691566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-catalog-content\") pod \"c622d8cc-2926-4e85-979b-37661d53380f\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.691878 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-utilities\") pod \"c622d8cc-2926-4e85-979b-37661d53380f\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.691999 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rzcf\" (UniqueName: \"kubernetes.io/projected/c622d8cc-2926-4e85-979b-37661d53380f-kube-api-access-6rzcf\") pod \"c622d8cc-2926-4e85-979b-37661d53380f\" (UID: \"c622d8cc-2926-4e85-979b-37661d53380f\") " Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.694896 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-utilities" (OuterVolumeSpecName: "utilities") pod "c622d8cc-2926-4e85-979b-37661d53380f" (UID: "c622d8cc-2926-4e85-979b-37661d53380f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.701843 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c622d8cc-2926-4e85-979b-37661d53380f-kube-api-access-6rzcf" (OuterVolumeSpecName: "kube-api-access-6rzcf") pod "c622d8cc-2926-4e85-979b-37661d53380f" (UID: "c622d8cc-2926-4e85-979b-37661d53380f"). InnerVolumeSpecName "kube-api-access-6rzcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.756610 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c622d8cc-2926-4e85-979b-37661d53380f" (UID: "c622d8cc-2926-4e85-979b-37661d53380f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.794787 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.794816 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rzcf\" (UniqueName: \"kubernetes.io/projected/c622d8cc-2926-4e85-979b-37661d53380f-kube-api-access-6rzcf\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:15 crc kubenswrapper[4926]: I1007 22:01:15.794827 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c622d8cc-2926-4e85-979b-37661d53380f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.517165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhxfc" event={"ID":"c622d8cc-2926-4e85-979b-37661d53380f","Type":"ContainerDied","Data":"54d00fc1e4f2578b09bd6e6ff1379c1b1edf2a7649b6deb5319dd6b9ddae5add"} Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.518341 4926 scope.go:117] "RemoveContainer" containerID="d75f39ba227a525d2cbfe3cc523665a988b76c030a7a0d79e632d15ce0914517" Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.517334 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhxfc" Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.578720 4926 scope.go:117] "RemoveContainer" containerID="891766edd3654cbd3e927447bd9d5deebbee4f9ec06526d03112d1bf133361ad" Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.598301 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vhxfc"] Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.612293 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vhxfc"] Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.621780 4926 scope.go:117] "RemoveContainer" containerID="6e79638c52e7b6ca402f6517a9cf9f3e997bb7d7b59f3822a42707850be77279" Oct 07 22:01:16 crc kubenswrapper[4926]: I1007 22:01:16.693299 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c622d8cc-2926-4e85-979b-37661d53380f" path="/var/lib/kubelet/pods/c622d8cc-2926-4e85-979b-37661d53380f/volumes" Oct 07 22:03:33 crc kubenswrapper[4926]: I1007 22:03:33.211030 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:03:33 crc kubenswrapper[4926]: I1007 22:03:33.211876 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:04:03 crc kubenswrapper[4926]: I1007 22:04:03.209994 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:04:03 crc kubenswrapper[4926]: I1007 22:04:03.212071 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.247769 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-84wmz"] Oct 07 22:04:14 crc kubenswrapper[4926]: E1007 22:04:14.248741 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="registry-server" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.248758 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="registry-server" Oct 07 22:04:14 crc kubenswrapper[4926]: E1007 22:04:14.248779 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4d73e0-30e0-4388-b7dc-28c97399930d" containerName="keystone-cron" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.248787 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4d73e0-30e0-4388-b7dc-28c97399930d" containerName="keystone-cron" Oct 07 22:04:14 crc kubenswrapper[4926]: E1007 22:04:14.248810 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="extract-utilities" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.248818 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="extract-utilities" Oct 07 22:04:14 crc kubenswrapper[4926]: E1007 22:04:14.248836 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="extract-content" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.248843 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="extract-content" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.249088 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4d73e0-30e0-4388-b7dc-28c97399930d" containerName="keystone-cron" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.249146 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c622d8cc-2926-4e85-979b-37661d53380f" containerName="registry-server" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.251055 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.263347 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84wmz"] Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.294831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-utilities\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.294976 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-catalog-content\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.295349 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfs94\" (UniqueName: \"kubernetes.io/projected/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-kube-api-access-vfs94\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.397723 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfs94\" (UniqueName: \"kubernetes.io/projected/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-kube-api-access-vfs94\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.397855 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-utilities\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.397902 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-catalog-content\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.398441 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-catalog-content\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.398558 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-utilities\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.416287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfs94\" (UniqueName: \"kubernetes.io/projected/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-kube-api-access-vfs94\") pod \"redhat-operators-84wmz\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:14 crc kubenswrapper[4926]: I1007 22:04:14.572440 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:15 crc kubenswrapper[4926]: I1007 22:04:15.066383 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84wmz"] Oct 07 22:04:15 crc kubenswrapper[4926]: W1007 22:04:15.069232 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4b4e14b_3fa7_458f_a1ec_fbdeb409d4cd.slice/crio-e342209e0c1433aa56e902e42a8cf2dc1235b79ddad35cc3ed03c9a081063c72 WatchSource:0}: Error finding container e342209e0c1433aa56e902e42a8cf2dc1235b79ddad35cc3ed03c9a081063c72: Status 404 returned error can't find the container with id e342209e0c1433aa56e902e42a8cf2dc1235b79ddad35cc3ed03c9a081063c72 Oct 07 22:04:15 crc kubenswrapper[4926]: I1007 22:04:15.508874 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerID="cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea" exitCode=0 Oct 07 22:04:15 crc kubenswrapper[4926]: I1007 22:04:15.508939 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84wmz" event={"ID":"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd","Type":"ContainerDied","Data":"cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea"} Oct 07 22:04:15 crc kubenswrapper[4926]: I1007 22:04:15.509126 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84wmz" event={"ID":"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd","Type":"ContainerStarted","Data":"e342209e0c1433aa56e902e42a8cf2dc1235b79ddad35cc3ed03c9a081063c72"} Oct 07 22:04:15 crc kubenswrapper[4926]: I1007 22:04:15.510598 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:04:20 crc kubenswrapper[4926]: I1007 22:04:20.575850 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerID="d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa" exitCode=0 Oct 07 22:04:20 crc kubenswrapper[4926]: I1007 22:04:20.575977 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84wmz" event={"ID":"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd","Type":"ContainerDied","Data":"d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa"} Oct 07 22:04:21 crc kubenswrapper[4926]: I1007 22:04:21.589278 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84wmz" event={"ID":"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd","Type":"ContainerStarted","Data":"6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a"} Oct 07 22:04:21 crc kubenswrapper[4926]: I1007 22:04:21.618496 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-84wmz" podStartSLOduration=1.919487119 podStartE2EDuration="7.618471973s" podCreationTimestamp="2025-10-07 22:04:14 +0000 UTC" firstStartedPulling="2025-10-07 22:04:15.510374241 +0000 UTC m=+4105.548955381" lastFinishedPulling="2025-10-07 22:04:21.209359085 +0000 UTC m=+4111.247940235" observedRunningTime="2025-10-07 22:04:21.604673683 +0000 UTC m=+4111.643254843" watchObservedRunningTime="2025-10-07 22:04:21.618471973 +0000 UTC m=+4111.657053143" Oct 07 22:04:24 crc kubenswrapper[4926]: I1007 22:04:24.573583 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:24 crc kubenswrapper[4926]: I1007 22:04:24.573998 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:26 crc kubenswrapper[4926]: I1007 22:04:26.130392 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-84wmz" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="registry-server" probeResult="failure" output=< Oct 07 22:04:26 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Oct 07 22:04:26 crc kubenswrapper[4926]: > Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.209301 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.210235 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.210309 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.211576 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"235edc795a0778f74569a034f6238258e3f846e8c83b30c0465bdc68d4b2383e"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.211674 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://235edc795a0778f74569a034f6238258e3f846e8c83b30c0465bdc68d4b2383e" gracePeriod=600 Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.728658 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="235edc795a0778f74569a034f6238258e3f846e8c83b30c0465bdc68d4b2383e" exitCode=0 Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.728878 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"235edc795a0778f74569a034f6238258e3f846e8c83b30c0465bdc68d4b2383e"} Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.729097 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2"} Oct 07 22:04:33 crc kubenswrapper[4926]: I1007 22:04:33.729123 4926 scope.go:117] "RemoveContainer" containerID="29662d5cbd8f1b82d24227901ee45d01c218668428fca413f85e6e8bb276a03d" Oct 07 22:04:34 crc kubenswrapper[4926]: I1007 22:04:34.674083 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:34 crc kubenswrapper[4926]: I1007 22:04:34.760416 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:34 crc kubenswrapper[4926]: I1007 22:04:34.925422 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84wmz"] Oct 07 22:04:35 crc kubenswrapper[4926]: I1007 22:04:35.757039 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-84wmz" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="registry-server" containerID="cri-o://6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a" gracePeriod=2 Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.250807 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.417483 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfs94\" (UniqueName: \"kubernetes.io/projected/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-kube-api-access-vfs94\") pod \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.417588 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-utilities\") pod \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.417675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-catalog-content\") pod \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\" (UID: \"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd\") " Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.424855 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-kube-api-access-vfs94" (OuterVolumeSpecName: "kube-api-access-vfs94") pod "b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" (UID: "b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd"). InnerVolumeSpecName "kube-api-access-vfs94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.426822 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-utilities" (OuterVolumeSpecName: "utilities") pod "b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" (UID: "b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.520364 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfs94\" (UniqueName: \"kubernetes.io/projected/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-kube-api-access-vfs94\") on node \"crc\" DevicePath \"\"" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.520412 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.554793 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" (UID: "b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.622340 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.771997 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerID="6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a" exitCode=0 Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.772076 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84wmz" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.772078 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84wmz" event={"ID":"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd","Type":"ContainerDied","Data":"6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a"} Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.772303 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84wmz" event={"ID":"b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd","Type":"ContainerDied","Data":"e342209e0c1433aa56e902e42a8cf2dc1235b79ddad35cc3ed03c9a081063c72"} Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.772379 4926 scope.go:117] "RemoveContainer" containerID="6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.811268 4926 scope.go:117] "RemoveContainer" containerID="d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.813147 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84wmz"] Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.823071 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-84wmz"] Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.837286 4926 scope.go:117] "RemoveContainer" containerID="cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.905975 4926 scope.go:117] "RemoveContainer" containerID="6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a" Oct 07 22:04:36 crc kubenswrapper[4926]: E1007 22:04:36.907020 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a\": container with ID starting with 6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a not found: ID does not exist" containerID="6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.907115 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a"} err="failed to get container status \"6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a\": rpc error: code = NotFound desc = could not find container \"6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a\": container with ID starting with 6db3d1e9c343e6c96a344c67cfa9cc88d15c049b473e8b39a04799f6adc4f49a not found: ID does not exist" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.907151 4926 scope.go:117] "RemoveContainer" containerID="d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa" Oct 07 22:04:36 crc kubenswrapper[4926]: E1007 22:04:36.907678 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa\": container with ID starting with d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa not found: ID does not exist" containerID="d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.907719 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa"} err="failed to get container status \"d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa\": rpc error: code = NotFound desc = could not find container \"d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa\": container with ID starting with d0a590585de90ced4c469f090306debb048870ff29faf2c38789a8d1ec50a3fa not found: ID does not exist" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.907747 4926 scope.go:117] "RemoveContainer" containerID="cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea" Oct 07 22:04:36 crc kubenswrapper[4926]: E1007 22:04:36.908348 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea\": container with ID starting with cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea not found: ID does not exist" containerID="cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea" Oct 07 22:04:36 crc kubenswrapper[4926]: I1007 22:04:36.908432 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea"} err="failed to get container status \"cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea\": rpc error: code = NotFound desc = could not find container \"cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea\": container with ID starting with cf5c8e43f952cd0016327b4ab185f6602abdd687275645cef76a73eca02c26ea not found: ID does not exist" Oct 07 22:04:38 crc kubenswrapper[4926]: I1007 22:04:38.695330 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" path="/var/lib/kubelet/pods/b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd/volumes" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.382990 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fkfw9"] Oct 07 22:06:26 crc kubenswrapper[4926]: E1007 22:06:26.384108 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="extract-utilities" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.384163 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="extract-utilities" Oct 07 22:06:26 crc kubenswrapper[4926]: E1007 22:06:26.384231 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="registry-server" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.384241 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="registry-server" Oct 07 22:06:26 crc kubenswrapper[4926]: E1007 22:06:26.384260 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="extract-content" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.384266 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="extract-content" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.384645 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b4e14b-3fa7-458f-a1ec-fbdeb409d4cd" containerName="registry-server" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.386602 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.398438 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkfw9"] Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.520154 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-utilities\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.520249 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrwrm\" (UniqueName: \"kubernetes.io/projected/956078b4-c026-48d7-b037-c11cea11c5c4-kube-api-access-hrwrm\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.520490 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-catalog-content\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.622541 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-catalog-content\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.622593 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-utilities\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.622625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrwrm\" (UniqueName: \"kubernetes.io/projected/956078b4-c026-48d7-b037-c11cea11c5c4-kube-api-access-hrwrm\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.623315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-utilities\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.623826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-catalog-content\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.648900 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrwrm\" (UniqueName: \"kubernetes.io/projected/956078b4-c026-48d7-b037-c11cea11c5c4-kube-api-access-hrwrm\") pod \"redhat-marketplace-fkfw9\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:26 crc kubenswrapper[4926]: I1007 22:06:26.718384 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:27 crc kubenswrapper[4926]: I1007 22:06:27.211495 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkfw9"] Oct 07 22:06:27 crc kubenswrapper[4926]: W1007 22:06:27.223426 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod956078b4_c026_48d7_b037_c11cea11c5c4.slice/crio-2e174751a006b84784cd0001af08f4aa5c5d9e00a0460643d450eef6effe8d49 WatchSource:0}: Error finding container 2e174751a006b84784cd0001af08f4aa5c5d9e00a0460643d450eef6effe8d49: Status 404 returned error can't find the container with id 2e174751a006b84784cd0001af08f4aa5c5d9e00a0460643d450eef6effe8d49 Oct 07 22:06:28 crc kubenswrapper[4926]: I1007 22:06:28.132046 4926 generic.go:334] "Generic (PLEG): container finished" podID="956078b4-c026-48d7-b037-c11cea11c5c4" containerID="9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9" exitCode=0 Oct 07 22:06:28 crc kubenswrapper[4926]: I1007 22:06:28.132112 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkfw9" event={"ID":"956078b4-c026-48d7-b037-c11cea11c5c4","Type":"ContainerDied","Data":"9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9"} Oct 07 22:06:28 crc kubenswrapper[4926]: I1007 22:06:28.132170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkfw9" event={"ID":"956078b4-c026-48d7-b037-c11cea11c5c4","Type":"ContainerStarted","Data":"2e174751a006b84784cd0001af08f4aa5c5d9e00a0460643d450eef6effe8d49"} Oct 07 22:06:30 crc kubenswrapper[4926]: I1007 22:06:30.165782 4926 generic.go:334] "Generic (PLEG): container finished" podID="956078b4-c026-48d7-b037-c11cea11c5c4" containerID="dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9" exitCode=0 Oct 07 22:06:30 crc kubenswrapper[4926]: I1007 22:06:30.165832 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkfw9" event={"ID":"956078b4-c026-48d7-b037-c11cea11c5c4","Type":"ContainerDied","Data":"dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9"} Oct 07 22:06:32 crc kubenswrapper[4926]: I1007 22:06:32.187979 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkfw9" event={"ID":"956078b4-c026-48d7-b037-c11cea11c5c4","Type":"ContainerStarted","Data":"b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864"} Oct 07 22:06:32 crc kubenswrapper[4926]: I1007 22:06:32.214806 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fkfw9" podStartSLOduration=3.6931922459999997 podStartE2EDuration="6.21478892s" podCreationTimestamp="2025-10-07 22:06:26 +0000 UTC" firstStartedPulling="2025-10-07 22:06:28.134522361 +0000 UTC m=+4238.173103521" lastFinishedPulling="2025-10-07 22:06:30.656119045 +0000 UTC m=+4240.694700195" observedRunningTime="2025-10-07 22:06:32.209836717 +0000 UTC m=+4242.248417887" watchObservedRunningTime="2025-10-07 22:06:32.21478892 +0000 UTC m=+4242.253370090" Oct 07 22:06:33 crc kubenswrapper[4926]: I1007 22:06:33.209077 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:06:33 crc kubenswrapper[4926]: I1007 22:06:33.209424 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:06:36 crc kubenswrapper[4926]: I1007 22:06:36.719243 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:36 crc kubenswrapper[4926]: I1007 22:06:36.720518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:36 crc kubenswrapper[4926]: I1007 22:06:36.771514 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:37 crc kubenswrapper[4926]: I1007 22:06:37.302914 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:37 crc kubenswrapper[4926]: I1007 22:06:37.360970 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkfw9"] Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.261940 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fkfw9" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="registry-server" containerID="cri-o://b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864" gracePeriod=2 Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.720274 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.839267 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrwrm\" (UniqueName: \"kubernetes.io/projected/956078b4-c026-48d7-b037-c11cea11c5c4-kube-api-access-hrwrm\") pod \"956078b4-c026-48d7-b037-c11cea11c5c4\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.839579 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-utilities\") pod \"956078b4-c026-48d7-b037-c11cea11c5c4\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.839641 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-catalog-content\") pod \"956078b4-c026-48d7-b037-c11cea11c5c4\" (UID: \"956078b4-c026-48d7-b037-c11cea11c5c4\") " Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.841098 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-utilities" (OuterVolumeSpecName: "utilities") pod "956078b4-c026-48d7-b037-c11cea11c5c4" (UID: "956078b4-c026-48d7-b037-c11cea11c5c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.850029 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/956078b4-c026-48d7-b037-c11cea11c5c4-kube-api-access-hrwrm" (OuterVolumeSpecName: "kube-api-access-hrwrm") pod "956078b4-c026-48d7-b037-c11cea11c5c4" (UID: "956078b4-c026-48d7-b037-c11cea11c5c4"). InnerVolumeSpecName "kube-api-access-hrwrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.853338 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "956078b4-c026-48d7-b037-c11cea11c5c4" (UID: "956078b4-c026-48d7-b037-c11cea11c5c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.942896 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrwrm\" (UniqueName: \"kubernetes.io/projected/956078b4-c026-48d7-b037-c11cea11c5c4-kube-api-access-hrwrm\") on node \"crc\" DevicePath \"\"" Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.942971 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:06:39 crc kubenswrapper[4926]: I1007 22:06:39.942999 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956078b4-c026-48d7-b037-c11cea11c5c4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.273062 4926 generic.go:334] "Generic (PLEG): container finished" podID="956078b4-c026-48d7-b037-c11cea11c5c4" containerID="b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864" exitCode=0 Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.273107 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkfw9" event={"ID":"956078b4-c026-48d7-b037-c11cea11c5c4","Type":"ContainerDied","Data":"b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864"} Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.273138 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkfw9" event={"ID":"956078b4-c026-48d7-b037-c11cea11c5c4","Type":"ContainerDied","Data":"2e174751a006b84784cd0001af08f4aa5c5d9e00a0460643d450eef6effe8d49"} Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.273158 4926 scope.go:117] "RemoveContainer" containerID="b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.273161 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkfw9" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.295844 4926 scope.go:117] "RemoveContainer" containerID="dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.321311 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkfw9"] Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.331758 4926 scope.go:117] "RemoveContainer" containerID="9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.332956 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkfw9"] Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.361644 4926 scope.go:117] "RemoveContainer" containerID="b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864" Oct 07 22:06:40 crc kubenswrapper[4926]: E1007 22:06:40.362048 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864\": container with ID starting with b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864 not found: ID does not exist" containerID="b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.362084 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864"} err="failed to get container status \"b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864\": rpc error: code = NotFound desc = could not find container \"b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864\": container with ID starting with b4e88ddcffda1efc6ed81e1243e774912a0635f25942de47bf555e26aeeae864 not found: ID does not exist" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.362112 4926 scope.go:117] "RemoveContainer" containerID="dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9" Oct 07 22:06:40 crc kubenswrapper[4926]: E1007 22:06:40.362554 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9\": container with ID starting with dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9 not found: ID does not exist" containerID="dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.362593 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9"} err="failed to get container status \"dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9\": rpc error: code = NotFound desc = could not find container \"dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9\": container with ID starting with dfc5b666d1b5d9b22adfa7720df4f9f56de5df5d593d2de1d89311becd6caaa9 not found: ID does not exist" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.362619 4926 scope.go:117] "RemoveContainer" containerID="9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9" Oct 07 22:06:40 crc kubenswrapper[4926]: E1007 22:06:40.362909 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9\": container with ID starting with 9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9 not found: ID does not exist" containerID="9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.362935 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9"} err="failed to get container status \"9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9\": rpc error: code = NotFound desc = could not find container \"9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9\": container with ID starting with 9ce0aac44b86370d88de667b25dd823b66d379b33e1296112527ffcf225968d9 not found: ID does not exist" Oct 07 22:06:40 crc kubenswrapper[4926]: I1007 22:06:40.690630 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" path="/var/lib/kubelet/pods/956078b4-c026-48d7-b037-c11cea11c5c4/volumes" Oct 07 22:07:03 crc kubenswrapper[4926]: I1007 22:07:03.210004 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:07:03 crc kubenswrapper[4926]: I1007 22:07:03.210448 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.209429 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.210029 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.210072 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.210879 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.210930 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" gracePeriod=600 Oct 07 22:07:33 crc kubenswrapper[4926]: E1007 22:07:33.334863 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.859127 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" exitCode=0 Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.859171 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2"} Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.859264 4926 scope.go:117] "RemoveContainer" containerID="235edc795a0778f74569a034f6238258e3f846e8c83b30c0465bdc68d4b2383e" Oct 07 22:07:33 crc kubenswrapper[4926]: I1007 22:07:33.860066 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:07:33 crc kubenswrapper[4926]: E1007 22:07:33.860669 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:07:47 crc kubenswrapper[4926]: I1007 22:07:47.679776 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:07:47 crc kubenswrapper[4926]: E1007 22:07:47.680676 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:07:59 crc kubenswrapper[4926]: I1007 22:07:59.679783 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:07:59 crc kubenswrapper[4926]: E1007 22:07:59.680491 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:08:12 crc kubenswrapper[4926]: I1007 22:08:12.679656 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:08:12 crc kubenswrapper[4926]: E1007 22:08:12.680796 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:08:27 crc kubenswrapper[4926]: I1007 22:08:27.679107 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:08:27 crc kubenswrapper[4926]: E1007 22:08:27.680057 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:08:39 crc kubenswrapper[4926]: I1007 22:08:39.681910 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:08:39 crc kubenswrapper[4926]: E1007 22:08:39.683064 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:08:53 crc kubenswrapper[4926]: I1007 22:08:53.679622 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:08:53 crc kubenswrapper[4926]: E1007 22:08:53.680983 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:09:05 crc kubenswrapper[4926]: I1007 22:09:05.679063 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:09:05 crc kubenswrapper[4926]: E1007 22:09:05.679945 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:09:17 crc kubenswrapper[4926]: I1007 22:09:17.679649 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:09:17 crc kubenswrapper[4926]: E1007 22:09:17.680637 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:09:29 crc kubenswrapper[4926]: I1007 22:09:29.679397 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:09:29 crc kubenswrapper[4926]: E1007 22:09:29.680529 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:09:44 crc kubenswrapper[4926]: I1007 22:09:44.679049 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:09:44 crc kubenswrapper[4926]: E1007 22:09:44.680146 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:09:58 crc kubenswrapper[4926]: I1007 22:09:58.679174 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:09:58 crc kubenswrapper[4926]: E1007 22:09:58.680531 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:10:02 crc kubenswrapper[4926]: E1007 22:10:02.978438 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.58:55732->38.102.83.58:42335: write tcp 38.102.83.58:55732->38.102.83.58:42335: write: broken pipe Oct 07 22:10:12 crc kubenswrapper[4926]: I1007 22:10:12.680222 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:10:12 crc kubenswrapper[4926]: E1007 22:10:12.681623 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.617766 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-76b2j"] Oct 07 22:10:20 crc kubenswrapper[4926]: E1007 22:10:20.619865 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="extract-content" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.619962 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="extract-content" Oct 07 22:10:20 crc kubenswrapper[4926]: E1007 22:10:20.620033 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="registry-server" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.620088 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="registry-server" Oct 07 22:10:20 crc kubenswrapper[4926]: E1007 22:10:20.620167 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="extract-utilities" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.620245 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="extract-utilities" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.620470 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="956078b4-c026-48d7-b037-c11cea11c5c4" containerName="registry-server" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.621965 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.638749 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76b2j"] Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.806267 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-utilities\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.806590 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-catalog-content\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.806755 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9dkk\" (UniqueName: \"kubernetes.io/projected/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-kube-api-access-h9dkk\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.908889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9dkk\" (UniqueName: \"kubernetes.io/projected/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-kube-api-access-h9dkk\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.909186 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-utilities\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.909321 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-catalog-content\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.909759 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-utilities\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.909782 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-catalog-content\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.935645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9dkk\" (UniqueName: \"kubernetes.io/projected/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-kube-api-access-h9dkk\") pod \"community-operators-76b2j\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:20 crc kubenswrapper[4926]: I1007 22:10:20.942570 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:21 crc kubenswrapper[4926]: I1007 22:10:21.491313 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76b2j"] Oct 07 22:10:21 crc kubenswrapper[4926]: I1007 22:10:21.785921 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerID="df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3" exitCode=0 Oct 07 22:10:21 crc kubenswrapper[4926]: I1007 22:10:21.785994 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76b2j" event={"ID":"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b","Type":"ContainerDied","Data":"df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3"} Oct 07 22:10:21 crc kubenswrapper[4926]: I1007 22:10:21.788290 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76b2j" event={"ID":"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b","Type":"ContainerStarted","Data":"0a17f4b0accaa0c0e3457517fac3f72145c1a1468756fa214573adea615de9c5"} Oct 07 22:10:21 crc kubenswrapper[4926]: I1007 22:10:21.788444 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:10:23 crc kubenswrapper[4926]: I1007 22:10:23.809775 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerID="724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202" exitCode=0 Oct 07 22:10:23 crc kubenswrapper[4926]: I1007 22:10:23.809862 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76b2j" event={"ID":"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b","Type":"ContainerDied","Data":"724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202"} Oct 07 22:10:24 crc kubenswrapper[4926]: I1007 22:10:24.825331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76b2j" event={"ID":"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b","Type":"ContainerStarted","Data":"abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6"} Oct 07 22:10:24 crc kubenswrapper[4926]: I1007 22:10:24.845246 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-76b2j" podStartSLOduration=2.344161133 podStartE2EDuration="4.845225972s" podCreationTimestamp="2025-10-07 22:10:20 +0000 UTC" firstStartedPulling="2025-10-07 22:10:21.787947102 +0000 UTC m=+4471.826528292" lastFinishedPulling="2025-10-07 22:10:24.289011941 +0000 UTC m=+4474.327593131" observedRunningTime="2025-10-07 22:10:24.840474154 +0000 UTC m=+4474.879055304" watchObservedRunningTime="2025-10-07 22:10:24.845225972 +0000 UTC m=+4474.883807122" Oct 07 22:10:26 crc kubenswrapper[4926]: I1007 22:10:26.679582 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:10:26 crc kubenswrapper[4926]: E1007 22:10:26.680645 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:10:30 crc kubenswrapper[4926]: I1007 22:10:30.943681 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:30 crc kubenswrapper[4926]: I1007 22:10:30.944187 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:31 crc kubenswrapper[4926]: I1007 22:10:31.000282 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:31 crc kubenswrapper[4926]: I1007 22:10:31.963518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:32 crc kubenswrapper[4926]: I1007 22:10:32.037969 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76b2j"] Oct 07 22:10:33 crc kubenswrapper[4926]: I1007 22:10:33.929848 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-76b2j" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="registry-server" containerID="cri-o://abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6" gracePeriod=2 Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.462678 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.601867 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9dkk\" (UniqueName: \"kubernetes.io/projected/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-kube-api-access-h9dkk\") pod \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.601998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-utilities\") pod \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.602042 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-catalog-content\") pod \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\" (UID: \"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b\") " Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.603594 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-utilities" (OuterVolumeSpecName: "utilities") pod "7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" (UID: "7b4f0fa2-c51f-4bdd-b908-86be72a94e2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.607279 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-kube-api-access-h9dkk" (OuterVolumeSpecName: "kube-api-access-h9dkk") pod "7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" (UID: "7b4f0fa2-c51f-4bdd-b908-86be72a94e2b"). InnerVolumeSpecName "kube-api-access-h9dkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.670688 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" (UID: "7b4f0fa2-c51f-4bdd-b908-86be72a94e2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.715086 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9dkk\" (UniqueName: \"kubernetes.io/projected/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-kube-api-access-h9dkk\") on node \"crc\" DevicePath \"\"" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.715132 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.715153 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.948043 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerID="abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6" exitCode=0 Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.948220 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76b2j" event={"ID":"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b","Type":"ContainerDied","Data":"abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6"} Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.948708 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76b2j" event={"ID":"7b4f0fa2-c51f-4bdd-b908-86be72a94e2b","Type":"ContainerDied","Data":"0a17f4b0accaa0c0e3457517fac3f72145c1a1468756fa214573adea615de9c5"} Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.948279 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76b2j" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.948751 4926 scope.go:117] "RemoveContainer" containerID="abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6" Oct 07 22:10:34 crc kubenswrapper[4926]: I1007 22:10:34.990275 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76b2j"] Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.004307 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-76b2j"] Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.013388 4926 scope.go:117] "RemoveContainer" containerID="724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.060622 4926 scope.go:117] "RemoveContainer" containerID="df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.116134 4926 scope.go:117] "RemoveContainer" containerID="abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6" Oct 07 22:10:35 crc kubenswrapper[4926]: E1007 22:10:35.116939 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6\": container with ID starting with abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6 not found: ID does not exist" containerID="abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.116986 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6"} err="failed to get container status \"abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6\": rpc error: code = NotFound desc = could not find container \"abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6\": container with ID starting with abaafd1720fc45f23f30a1b66b50c4e3c73a17f3f6e182a8654f12fba1c124c6 not found: ID does not exist" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.117019 4926 scope.go:117] "RemoveContainer" containerID="724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202" Oct 07 22:10:35 crc kubenswrapper[4926]: E1007 22:10:35.117533 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202\": container with ID starting with 724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202 not found: ID does not exist" containerID="724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.117593 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202"} err="failed to get container status \"724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202\": rpc error: code = NotFound desc = could not find container \"724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202\": container with ID starting with 724c5d5b96ea635814b5292e29a672b0f0e4eb8c89488cb916ea7c95a0d26202 not found: ID does not exist" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.117633 4926 scope.go:117] "RemoveContainer" containerID="df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3" Oct 07 22:10:35 crc kubenswrapper[4926]: E1007 22:10:35.117955 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3\": container with ID starting with df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3 not found: ID does not exist" containerID="df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3" Oct 07 22:10:35 crc kubenswrapper[4926]: I1007 22:10:35.117992 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3"} err="failed to get container status \"df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3\": rpc error: code = NotFound desc = could not find container \"df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3\": container with ID starting with df566a7b202e4876b655916923306c31e4f9e210616678cf559df582252e7ef3 not found: ID does not exist" Oct 07 22:10:36 crc kubenswrapper[4926]: I1007 22:10:36.698493 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" path="/var/lib/kubelet/pods/7b4f0fa2-c51f-4bdd-b908-86be72a94e2b/volumes" Oct 07 22:10:38 crc kubenswrapper[4926]: I1007 22:10:38.680001 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:10:38 crc kubenswrapper[4926]: E1007 22:10:38.680984 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:10:49 crc kubenswrapper[4926]: I1007 22:10:49.679320 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:10:49 crc kubenswrapper[4926]: E1007 22:10:49.680092 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:11:00 crc kubenswrapper[4926]: I1007 22:11:00.685780 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:11:00 crc kubenswrapper[4926]: E1007 22:11:00.686904 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:11:12 crc kubenswrapper[4926]: I1007 22:11:12.679609 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:11:12 crc kubenswrapper[4926]: E1007 22:11:12.680816 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:11:25 crc kubenswrapper[4926]: I1007 22:11:25.679107 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:11:25 crc kubenswrapper[4926]: E1007 22:11:25.680130 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:11:39 crc kubenswrapper[4926]: I1007 22:11:39.679086 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:11:39 crc kubenswrapper[4926]: E1007 22:11:39.679788 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:11:51 crc kubenswrapper[4926]: I1007 22:11:51.679460 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:11:51 crc kubenswrapper[4926]: E1007 22:11:51.680229 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.928056 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kq2pr"] Oct 07 22:12:00 crc kubenswrapper[4926]: E1007 22:12:00.929868 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="registry-server" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.929887 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="registry-server" Oct 07 22:12:00 crc kubenswrapper[4926]: E1007 22:12:00.929915 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="extract-content" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.929923 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="extract-content" Oct 07 22:12:00 crc kubenswrapper[4926]: E1007 22:12:00.929934 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="extract-utilities" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.929946 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="extract-utilities" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.930226 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4f0fa2-c51f-4bdd-b908-86be72a94e2b" containerName="registry-server" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.931901 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:00 crc kubenswrapper[4926]: I1007 22:12:00.949949 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kq2pr"] Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.065921 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-utilities\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.066364 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-catalog-content\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.066423 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxwjb\" (UniqueName: \"kubernetes.io/projected/99856a53-57b4-4681-8482-4c85f849281c-kube-api-access-xxwjb\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.169043 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-catalog-content\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.169113 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxwjb\" (UniqueName: \"kubernetes.io/projected/99856a53-57b4-4681-8482-4c85f849281c-kube-api-access-xxwjb\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.169377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-utilities\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.169926 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-catalog-content\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.170104 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-utilities\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.194480 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxwjb\" (UniqueName: \"kubernetes.io/projected/99856a53-57b4-4681-8482-4c85f849281c-kube-api-access-xxwjb\") pod \"certified-operators-kq2pr\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.258633 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.804456 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kq2pr"] Oct 07 22:12:01 crc kubenswrapper[4926]: I1007 22:12:01.904854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kq2pr" event={"ID":"99856a53-57b4-4681-8482-4c85f849281c","Type":"ContainerStarted","Data":"b64736fb2cbf76c63632701b79a21eb0f34343c762f979346d918749e6823f91"} Oct 07 22:12:02 crc kubenswrapper[4926]: I1007 22:12:02.921362 4926 generic.go:334] "Generic (PLEG): container finished" podID="99856a53-57b4-4681-8482-4c85f849281c" containerID="5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c" exitCode=0 Oct 07 22:12:02 crc kubenswrapper[4926]: I1007 22:12:02.921546 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kq2pr" event={"ID":"99856a53-57b4-4681-8482-4c85f849281c","Type":"ContainerDied","Data":"5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c"} Oct 07 22:12:03 crc kubenswrapper[4926]: I1007 22:12:03.679545 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:12:03 crc kubenswrapper[4926]: E1007 22:12:03.680188 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:12:04 crc kubenswrapper[4926]: I1007 22:12:04.951692 4926 generic.go:334] "Generic (PLEG): container finished" podID="99856a53-57b4-4681-8482-4c85f849281c" containerID="4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa" exitCode=0 Oct 07 22:12:04 crc kubenswrapper[4926]: I1007 22:12:04.951751 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kq2pr" event={"ID":"99856a53-57b4-4681-8482-4c85f849281c","Type":"ContainerDied","Data":"4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa"} Oct 07 22:12:05 crc kubenswrapper[4926]: I1007 22:12:05.963982 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kq2pr" event={"ID":"99856a53-57b4-4681-8482-4c85f849281c","Type":"ContainerStarted","Data":"f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e"} Oct 07 22:12:06 crc kubenswrapper[4926]: I1007 22:12:06.014270 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kq2pr" podStartSLOduration=3.49325466 podStartE2EDuration="6.014236737s" podCreationTimestamp="2025-10-07 22:12:00 +0000 UTC" firstStartedPulling="2025-10-07 22:12:02.924768213 +0000 UTC m=+4572.963349403" lastFinishedPulling="2025-10-07 22:12:05.44575032 +0000 UTC m=+4575.484331480" observedRunningTime="2025-10-07 22:12:05.991784016 +0000 UTC m=+4576.030365176" watchObservedRunningTime="2025-10-07 22:12:06.014236737 +0000 UTC m=+4576.052818107" Oct 07 22:12:11 crc kubenswrapper[4926]: I1007 22:12:11.259605 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:11 crc kubenswrapper[4926]: I1007 22:12:11.259900 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:11 crc kubenswrapper[4926]: I1007 22:12:11.363993 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:12 crc kubenswrapper[4926]: I1007 22:12:12.095598 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:12 crc kubenswrapper[4926]: I1007 22:12:12.166600 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kq2pr"] Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.044279 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kq2pr" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="registry-server" containerID="cri-o://f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e" gracePeriod=2 Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.623602 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.786231 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxwjb\" (UniqueName: \"kubernetes.io/projected/99856a53-57b4-4681-8482-4c85f849281c-kube-api-access-xxwjb\") pod \"99856a53-57b4-4681-8482-4c85f849281c\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.786309 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-utilities\") pod \"99856a53-57b4-4681-8482-4c85f849281c\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.786512 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-catalog-content\") pod \"99856a53-57b4-4681-8482-4c85f849281c\" (UID: \"99856a53-57b4-4681-8482-4c85f849281c\") " Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.788242 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-utilities" (OuterVolumeSpecName: "utilities") pod "99856a53-57b4-4681-8482-4c85f849281c" (UID: "99856a53-57b4-4681-8482-4c85f849281c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.797418 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99856a53-57b4-4681-8482-4c85f849281c-kube-api-access-xxwjb" (OuterVolumeSpecName: "kube-api-access-xxwjb") pod "99856a53-57b4-4681-8482-4c85f849281c" (UID: "99856a53-57b4-4681-8482-4c85f849281c"). InnerVolumeSpecName "kube-api-access-xxwjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.836220 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99856a53-57b4-4681-8482-4c85f849281c" (UID: "99856a53-57b4-4681-8482-4c85f849281c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.888682 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.888730 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99856a53-57b4-4681-8482-4c85f849281c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:14 crc kubenswrapper[4926]: I1007 22:12:14.888748 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxwjb\" (UniqueName: \"kubernetes.io/projected/99856a53-57b4-4681-8482-4c85f849281c-kube-api-access-xxwjb\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.060043 4926 generic.go:334] "Generic (PLEG): container finished" podID="99856a53-57b4-4681-8482-4c85f849281c" containerID="f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e" exitCode=0 Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.060093 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kq2pr" event={"ID":"99856a53-57b4-4681-8482-4c85f849281c","Type":"ContainerDied","Data":"f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e"} Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.060119 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kq2pr" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.060141 4926 scope.go:117] "RemoveContainer" containerID="f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.060124 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kq2pr" event={"ID":"99856a53-57b4-4681-8482-4c85f849281c","Type":"ContainerDied","Data":"b64736fb2cbf76c63632701b79a21eb0f34343c762f979346d918749e6823f91"} Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.097690 4926 scope.go:117] "RemoveContainer" containerID="4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.142258 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kq2pr"] Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.152966 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kq2pr"] Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.154813 4926 scope.go:117] "RemoveContainer" containerID="5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.205336 4926 scope.go:117] "RemoveContainer" containerID="f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e" Oct 07 22:12:15 crc kubenswrapper[4926]: E1007 22:12:15.205938 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e\": container with ID starting with f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e not found: ID does not exist" containerID="f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.205998 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e"} err="failed to get container status \"f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e\": rpc error: code = NotFound desc = could not find container \"f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e\": container with ID starting with f5e368866ddca8f17d89d302a55a7e911c6478fe7d2ea6fb778183cab172309e not found: ID does not exist" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.206030 4926 scope.go:117] "RemoveContainer" containerID="4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa" Oct 07 22:12:15 crc kubenswrapper[4926]: E1007 22:12:15.206573 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa\": container with ID starting with 4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa not found: ID does not exist" containerID="4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.206607 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa"} err="failed to get container status \"4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa\": rpc error: code = NotFound desc = could not find container \"4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa\": container with ID starting with 4034a456ee3a4f035aec2e5ee55a8581cedbea1ef652a1a95a27544cbf96cfaa not found: ID does not exist" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.206627 4926 scope.go:117] "RemoveContainer" containerID="5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c" Oct 07 22:12:15 crc kubenswrapper[4926]: E1007 22:12:15.207028 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c\": container with ID starting with 5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c not found: ID does not exist" containerID="5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.207068 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c"} err="failed to get container status \"5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c\": rpc error: code = NotFound desc = could not find container \"5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c\": container with ID starting with 5042b09348f6d066a0e69a1041f283df3a416594e0272275f0c246fba6ecf99c not found: ID does not exist" Oct 07 22:12:15 crc kubenswrapper[4926]: I1007 22:12:15.679786 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:12:15 crc kubenswrapper[4926]: E1007 22:12:15.681024 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:12:16 crc kubenswrapper[4926]: I1007 22:12:16.699304 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99856a53-57b4-4681-8482-4c85f849281c" path="/var/lib/kubelet/pods/99856a53-57b4-4681-8482-4c85f849281c/volumes" Oct 07 22:12:28 crc kubenswrapper[4926]: I1007 22:12:28.678909 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:12:28 crc kubenswrapper[4926]: E1007 22:12:28.679549 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:12:39 crc kubenswrapper[4926]: I1007 22:12:39.679044 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:12:40 crc kubenswrapper[4926]: I1007 22:12:40.355789 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"fdb91b1da5a8ebd47cd24c5600bd185da1e7c9758810f66461048b5b9a04306f"} Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.159042 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75"] Oct 07 22:15:00 crc kubenswrapper[4926]: E1007 22:15:00.160640 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="extract-utilities" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.160661 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="extract-utilities" Oct 07 22:15:00 crc kubenswrapper[4926]: E1007 22:15:00.160699 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="extract-content" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.160707 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="extract-content" Oct 07 22:15:00 crc kubenswrapper[4926]: E1007 22:15:00.160729 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="registry-server" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.160738 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="registry-server" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.161051 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="99856a53-57b4-4681-8482-4c85f849281c" containerName="registry-server" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.161994 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.165622 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.165926 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.186566 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75"] Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.261530 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7529028c-41d2-48d1-a65a-a2f273541b5d-config-volume\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.261570 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfn9h\" (UniqueName: \"kubernetes.io/projected/7529028c-41d2-48d1-a65a-a2f273541b5d-kube-api-access-qfn9h\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.261646 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7529028c-41d2-48d1-a65a-a2f273541b5d-secret-volume\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.363735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7529028c-41d2-48d1-a65a-a2f273541b5d-config-volume\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.364191 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfn9h\" (UniqueName: \"kubernetes.io/projected/7529028c-41d2-48d1-a65a-a2f273541b5d-kube-api-access-qfn9h\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.364403 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7529028c-41d2-48d1-a65a-a2f273541b5d-secret-volume\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.365375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7529028c-41d2-48d1-a65a-a2f273541b5d-config-volume\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.375820 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7529028c-41d2-48d1-a65a-a2f273541b5d-secret-volume\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.393882 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfn9h\" (UniqueName: \"kubernetes.io/projected/7529028c-41d2-48d1-a65a-a2f273541b5d-kube-api-access-qfn9h\") pod \"collect-profiles-29331255-75b75\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.499285 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:00 crc kubenswrapper[4926]: I1007 22:15:00.945768 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75"] Oct 07 22:15:02 crc kubenswrapper[4926]: I1007 22:15:02.009453 4926 generic.go:334] "Generic (PLEG): container finished" podID="7529028c-41d2-48d1-a65a-a2f273541b5d" containerID="2276af94ad17dac1731d5883f11c1bbfc1e70af3d8b9622745f9a1046a13bb5c" exitCode=0 Oct 07 22:15:02 crc kubenswrapper[4926]: I1007 22:15:02.009607 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" event={"ID":"7529028c-41d2-48d1-a65a-a2f273541b5d","Type":"ContainerDied","Data":"2276af94ad17dac1731d5883f11c1bbfc1e70af3d8b9622745f9a1046a13bb5c"} Oct 07 22:15:02 crc kubenswrapper[4926]: I1007 22:15:02.009832 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" event={"ID":"7529028c-41d2-48d1-a65a-a2f273541b5d","Type":"ContainerStarted","Data":"d0a81e32f458dd754090ad45a9e191af2a1550017b3626bfd26a021fcb2a7081"} Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.211571 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.212035 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.433369 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.534515 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7529028c-41d2-48d1-a65a-a2f273541b5d-secret-volume\") pod \"7529028c-41d2-48d1-a65a-a2f273541b5d\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.534637 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7529028c-41d2-48d1-a65a-a2f273541b5d-config-volume\") pod \"7529028c-41d2-48d1-a65a-a2f273541b5d\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.534815 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfn9h\" (UniqueName: \"kubernetes.io/projected/7529028c-41d2-48d1-a65a-a2f273541b5d-kube-api-access-qfn9h\") pod \"7529028c-41d2-48d1-a65a-a2f273541b5d\" (UID: \"7529028c-41d2-48d1-a65a-a2f273541b5d\") " Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.535280 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7529028c-41d2-48d1-a65a-a2f273541b5d-config-volume" (OuterVolumeSpecName: "config-volume") pod "7529028c-41d2-48d1-a65a-a2f273541b5d" (UID: "7529028c-41d2-48d1-a65a-a2f273541b5d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.556312 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7529028c-41d2-48d1-a65a-a2f273541b5d-kube-api-access-qfn9h" (OuterVolumeSpecName: "kube-api-access-qfn9h") pod "7529028c-41d2-48d1-a65a-a2f273541b5d" (UID: "7529028c-41d2-48d1-a65a-a2f273541b5d"). InnerVolumeSpecName "kube-api-access-qfn9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.557172 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7529028c-41d2-48d1-a65a-a2f273541b5d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7529028c-41d2-48d1-a65a-a2f273541b5d" (UID: "7529028c-41d2-48d1-a65a-a2f273541b5d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.638242 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfn9h\" (UniqueName: \"kubernetes.io/projected/7529028c-41d2-48d1-a65a-a2f273541b5d-kube-api-access-qfn9h\") on node \"crc\" DevicePath \"\"" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.638282 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7529028c-41d2-48d1-a65a-a2f273541b5d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:15:03 crc kubenswrapper[4926]: I1007 22:15:03.638298 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7529028c-41d2-48d1-a65a-a2f273541b5d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:15:04 crc kubenswrapper[4926]: I1007 22:15:04.029277 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" event={"ID":"7529028c-41d2-48d1-a65a-a2f273541b5d","Type":"ContainerDied","Data":"d0a81e32f458dd754090ad45a9e191af2a1550017b3626bfd26a021fcb2a7081"} Oct 07 22:15:04 crc kubenswrapper[4926]: I1007 22:15:04.029351 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0a81e32f458dd754090ad45a9e191af2a1550017b3626bfd26a021fcb2a7081" Oct 07 22:15:04 crc kubenswrapper[4926]: I1007 22:15:04.029293 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-75b75" Oct 07 22:15:04 crc kubenswrapper[4926]: I1007 22:15:04.526236 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs"] Oct 07 22:15:04 crc kubenswrapper[4926]: I1007 22:15:04.538491 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331210-bphvs"] Oct 07 22:15:04 crc kubenswrapper[4926]: I1007 22:15:04.698057 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c513138c-31c6-4b48-94c1-71660c5af953" path="/var/lib/kubelet/pods/c513138c-31c6-4b48-94c1-71660c5af953/volumes" Oct 07 22:15:06 crc kubenswrapper[4926]: I1007 22:15:06.415668 4926 scope.go:117] "RemoveContainer" containerID="d6644ae4256546b50892aae953071a6ef593bb9df8619e18928756fdef1d6d06" Oct 07 22:15:33 crc kubenswrapper[4926]: I1007 22:15:33.209656 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:15:33 crc kubenswrapper[4926]: I1007 22:15:33.210325 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.209578 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.210221 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.210284 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.211174 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdb91b1da5a8ebd47cd24c5600bd185da1e7c9758810f66461048b5b9a04306f"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.211263 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://fdb91b1da5a8ebd47cd24c5600bd185da1e7c9758810f66461048b5b9a04306f" gracePeriod=600 Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.758126 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="fdb91b1da5a8ebd47cd24c5600bd185da1e7c9758810f66461048b5b9a04306f" exitCode=0 Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.758180 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"fdb91b1da5a8ebd47cd24c5600bd185da1e7c9758810f66461048b5b9a04306f"} Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.758645 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58"} Oct 07 22:16:03 crc kubenswrapper[4926]: I1007 22:16:03.758703 4926 scope.go:117] "RemoveContainer" containerID="93b148f414220ee13ab5aeeb403cc25590106075adf51c14fb81092e2ab622f2" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.002679 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8xkrj"] Oct 07 22:17:24 crc kubenswrapper[4926]: E1007 22:17:24.003516 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7529028c-41d2-48d1-a65a-a2f273541b5d" containerName="collect-profiles" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.003528 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7529028c-41d2-48d1-a65a-a2f273541b5d" containerName="collect-profiles" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.003744 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7529028c-41d2-48d1-a65a-a2f273541b5d" containerName="collect-profiles" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.005330 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.028209 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xkrj"] Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.163643 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrlqz\" (UniqueName: \"kubernetes.io/projected/65f76154-3c6f-4471-aea0-5d660fb9ce6b-kube-api-access-rrlqz\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.164086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-catalog-content\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.164118 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-utilities\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.266122 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrlqz\" (UniqueName: \"kubernetes.io/projected/65f76154-3c6f-4471-aea0-5d660fb9ce6b-kube-api-access-rrlqz\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.266359 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-catalog-content\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.266387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-utilities\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.266876 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-utilities\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.267065 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-catalog-content\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.473272 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrlqz\" (UniqueName: \"kubernetes.io/projected/65f76154-3c6f-4471-aea0-5d660fb9ce6b-kube-api-access-rrlqz\") pod \"redhat-marketplace-8xkrj\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:24 crc kubenswrapper[4926]: I1007 22:17:24.635205 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:25 crc kubenswrapper[4926]: I1007 22:17:25.154570 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xkrj"] Oct 07 22:17:25 crc kubenswrapper[4926]: I1007 22:17:25.666105 4926 generic.go:334] "Generic (PLEG): container finished" podID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerID="a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15" exitCode=0 Oct 07 22:17:25 crc kubenswrapper[4926]: I1007 22:17:25.666223 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xkrj" event={"ID":"65f76154-3c6f-4471-aea0-5d660fb9ce6b","Type":"ContainerDied","Data":"a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15"} Oct 07 22:17:25 crc kubenswrapper[4926]: I1007 22:17:25.666456 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xkrj" event={"ID":"65f76154-3c6f-4471-aea0-5d660fb9ce6b","Type":"ContainerStarted","Data":"0dfb29d443ec99364945d68d43aa35b8090fff2a090f048ae69d24bab27e6bdf"} Oct 07 22:17:25 crc kubenswrapper[4926]: I1007 22:17:25.668276 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:17:27 crc kubenswrapper[4926]: I1007 22:17:27.688671 4926 generic.go:334] "Generic (PLEG): container finished" podID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerID="d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543" exitCode=0 Oct 07 22:17:27 crc kubenswrapper[4926]: I1007 22:17:27.688748 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xkrj" event={"ID":"65f76154-3c6f-4471-aea0-5d660fb9ce6b","Type":"ContainerDied","Data":"d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543"} Oct 07 22:17:29 crc kubenswrapper[4926]: I1007 22:17:29.715271 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xkrj" event={"ID":"65f76154-3c6f-4471-aea0-5d660fb9ce6b","Type":"ContainerStarted","Data":"9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95"} Oct 07 22:17:29 crc kubenswrapper[4926]: I1007 22:17:29.751386 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8xkrj" podStartSLOduration=3.518161085 podStartE2EDuration="6.751353403s" podCreationTimestamp="2025-10-07 22:17:23 +0000 UTC" firstStartedPulling="2025-10-07 22:17:25.668052796 +0000 UTC m=+4895.706633946" lastFinishedPulling="2025-10-07 22:17:28.901245114 +0000 UTC m=+4898.939826264" observedRunningTime="2025-10-07 22:17:29.735456212 +0000 UTC m=+4899.774037372" watchObservedRunningTime="2025-10-07 22:17:29.751353403 +0000 UTC m=+4899.789934583" Oct 07 22:17:34 crc kubenswrapper[4926]: I1007 22:17:34.635949 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:34 crc kubenswrapper[4926]: I1007 22:17:34.636677 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:34 crc kubenswrapper[4926]: I1007 22:17:34.724807 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:34 crc kubenswrapper[4926]: I1007 22:17:34.818611 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:34 crc kubenswrapper[4926]: I1007 22:17:34.958117 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xkrj"] Oct 07 22:17:36 crc kubenswrapper[4926]: I1007 22:17:36.788283 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8xkrj" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="registry-server" containerID="cri-o://9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95" gracePeriod=2 Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.298573 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.457784 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-catalog-content\") pod \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.457877 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-utilities\") pod \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.457964 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrlqz\" (UniqueName: \"kubernetes.io/projected/65f76154-3c6f-4471-aea0-5d660fb9ce6b-kube-api-access-rrlqz\") pod \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\" (UID: \"65f76154-3c6f-4471-aea0-5d660fb9ce6b\") " Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.458988 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-utilities" (OuterVolumeSpecName: "utilities") pod "65f76154-3c6f-4471-aea0-5d660fb9ce6b" (UID: "65f76154-3c6f-4471-aea0-5d660fb9ce6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.465481 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f76154-3c6f-4471-aea0-5d660fb9ce6b-kube-api-access-rrlqz" (OuterVolumeSpecName: "kube-api-access-rrlqz") pod "65f76154-3c6f-4471-aea0-5d660fb9ce6b" (UID: "65f76154-3c6f-4471-aea0-5d660fb9ce6b"). InnerVolumeSpecName "kube-api-access-rrlqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.494172 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65f76154-3c6f-4471-aea0-5d660fb9ce6b" (UID: "65f76154-3c6f-4471-aea0-5d660fb9ce6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.561452 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.561511 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrlqz\" (UniqueName: \"kubernetes.io/projected/65f76154-3c6f-4471-aea0-5d660fb9ce6b-kube-api-access-rrlqz\") on node \"crc\" DevicePath \"\"" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.561527 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65f76154-3c6f-4471-aea0-5d660fb9ce6b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.800448 4926 generic.go:334] "Generic (PLEG): container finished" podID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerID="9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95" exitCode=0 Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.800500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xkrj" event={"ID":"65f76154-3c6f-4471-aea0-5d660fb9ce6b","Type":"ContainerDied","Data":"9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95"} Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.800531 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xkrj" event={"ID":"65f76154-3c6f-4471-aea0-5d660fb9ce6b","Type":"ContainerDied","Data":"0dfb29d443ec99364945d68d43aa35b8090fff2a090f048ae69d24bab27e6bdf"} Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.800547 4926 scope.go:117] "RemoveContainer" containerID="9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.800542 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xkrj" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.832205 4926 scope.go:117] "RemoveContainer" containerID="d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.839259 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xkrj"] Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.847054 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xkrj"] Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.866312 4926 scope.go:117] "RemoveContainer" containerID="a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.913557 4926 scope.go:117] "RemoveContainer" containerID="9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95" Oct 07 22:17:37 crc kubenswrapper[4926]: E1007 22:17:37.913940 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95\": container with ID starting with 9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95 not found: ID does not exist" containerID="9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.913983 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95"} err="failed to get container status \"9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95\": rpc error: code = NotFound desc = could not find container \"9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95\": container with ID starting with 9ac6b19ce60bb333de08cff4cae5bde78f9c62ad9505701ea68c12abd9e8dc95 not found: ID does not exist" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.914008 4926 scope.go:117] "RemoveContainer" containerID="d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543" Oct 07 22:17:37 crc kubenswrapper[4926]: E1007 22:17:37.914326 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543\": container with ID starting with d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543 not found: ID does not exist" containerID="d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.914360 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543"} err="failed to get container status \"d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543\": rpc error: code = NotFound desc = could not find container \"d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543\": container with ID starting with d30c5c44f470b2dbdb1c989a09a35ce5751a691073fe1894e50428d8b84b6543 not found: ID does not exist" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.914381 4926 scope.go:117] "RemoveContainer" containerID="a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15" Oct 07 22:17:37 crc kubenswrapper[4926]: E1007 22:17:37.914656 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15\": container with ID starting with a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15 not found: ID does not exist" containerID="a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15" Oct 07 22:17:37 crc kubenswrapper[4926]: I1007 22:17:37.914688 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15"} err="failed to get container status \"a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15\": rpc error: code = NotFound desc = could not find container \"a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15\": container with ID starting with a1d253d47bf12e66d7a675dd3477f513cf79ef335aedcbaa133c38d519f1bb15 not found: ID does not exist" Oct 07 22:17:38 crc kubenswrapper[4926]: I1007 22:17:38.690779 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" path="/var/lib/kubelet/pods/65f76154-3c6f-4471-aea0-5d660fb9ce6b/volumes" Oct 07 22:18:03 crc kubenswrapper[4926]: I1007 22:18:03.209387 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:18:03 crc kubenswrapper[4926]: I1007 22:18:03.210157 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:18:33 crc kubenswrapper[4926]: I1007 22:18:33.209439 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:18:33 crc kubenswrapper[4926]: I1007 22:18:33.209915 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:19:03 crc kubenswrapper[4926]: I1007 22:19:03.209153 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:19:03 crc kubenswrapper[4926]: I1007 22:19:03.209930 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:19:03 crc kubenswrapper[4926]: I1007 22:19:03.210014 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:19:03 crc kubenswrapper[4926]: I1007 22:19:03.211314 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:19:03 crc kubenswrapper[4926]: I1007 22:19:03.211449 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" gracePeriod=600 Oct 07 22:19:04 crc kubenswrapper[4926]: E1007 22:19:04.201100 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:19:04 crc kubenswrapper[4926]: I1007 22:19:04.737922 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" exitCode=0 Oct 07 22:19:04 crc kubenswrapper[4926]: I1007 22:19:04.738029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58"} Oct 07 22:19:04 crc kubenswrapper[4926]: I1007 22:19:04.738455 4926 scope.go:117] "RemoveContainer" containerID="fdb91b1da5a8ebd47cd24c5600bd185da1e7c9758810f66461048b5b9a04306f" Oct 07 22:19:04 crc kubenswrapper[4926]: I1007 22:19:04.739121 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:19:04 crc kubenswrapper[4926]: E1007 22:19:04.739552 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:19:18 crc kubenswrapper[4926]: I1007 22:19:18.679474 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:19:18 crc kubenswrapper[4926]: E1007 22:19:18.680801 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:19:32 crc kubenswrapper[4926]: I1007 22:19:32.679643 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:19:32 crc kubenswrapper[4926]: E1007 22:19:32.680486 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:19:44 crc kubenswrapper[4926]: I1007 22:19:44.678815 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:19:44 crc kubenswrapper[4926]: E1007 22:19:44.679617 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.679005 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:19:58 crc kubenswrapper[4926]: E1007 22:19:58.681565 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.748785 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wsmft"] Oct 07 22:19:58 crc kubenswrapper[4926]: E1007 22:19:58.749502 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="extract-content" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.749526 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="extract-content" Oct 07 22:19:58 crc kubenswrapper[4926]: E1007 22:19:58.749539 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="extract-utilities" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.749548 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="extract-utilities" Oct 07 22:19:58 crc kubenswrapper[4926]: E1007 22:19:58.749604 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="registry-server" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.749612 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="registry-server" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.749940 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f76154-3c6f-4471-aea0-5d660fb9ce6b" containerName="registry-server" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.751992 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.763670 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsmft"] Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.871707 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-catalog-content\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.871968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-utilities\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.872044 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r4vg\" (UniqueName: \"kubernetes.io/projected/a84b8ddc-1a7c-4e19-8320-6944446a3573-kube-api-access-4r4vg\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.974023 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r4vg\" (UniqueName: \"kubernetes.io/projected/a84b8ddc-1a7c-4e19-8320-6944446a3573-kube-api-access-4r4vg\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.974365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-catalog-content\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.974627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-utilities\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.975152 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-catalog-content\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.975176 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-utilities\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:58 crc kubenswrapper[4926]: I1007 22:19:58.996320 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r4vg\" (UniqueName: \"kubernetes.io/projected/a84b8ddc-1a7c-4e19-8320-6944446a3573-kube-api-access-4r4vg\") pod \"redhat-operators-wsmft\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:59 crc kubenswrapper[4926]: I1007 22:19:59.087626 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:19:59 crc kubenswrapper[4926]: I1007 22:19:59.550862 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsmft"] Oct 07 22:20:00 crc kubenswrapper[4926]: I1007 22:20:00.392288 4926 generic.go:334] "Generic (PLEG): container finished" podID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerID="5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc" exitCode=0 Oct 07 22:20:00 crc kubenswrapper[4926]: I1007 22:20:00.392794 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerDied","Data":"5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc"} Oct 07 22:20:00 crc kubenswrapper[4926]: I1007 22:20:00.392837 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerStarted","Data":"1ff486d872468ae08c3d1dd1709001fd113ad2509249dfa114fec94d29e999b3"} Oct 07 22:20:03 crc kubenswrapper[4926]: I1007 22:20:03.441329 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerStarted","Data":"6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611"} Oct 07 22:20:05 crc kubenswrapper[4926]: I1007 22:20:05.467150 4926 generic.go:334] "Generic (PLEG): container finished" podID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerID="6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611" exitCode=0 Oct 07 22:20:05 crc kubenswrapper[4926]: I1007 22:20:05.467252 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerDied","Data":"6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611"} Oct 07 22:20:07 crc kubenswrapper[4926]: I1007 22:20:07.491804 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerStarted","Data":"0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2"} Oct 07 22:20:07 crc kubenswrapper[4926]: I1007 22:20:07.522813 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wsmft" podStartSLOduration=3.489478849 podStartE2EDuration="9.522791994s" podCreationTimestamp="2025-10-07 22:19:58 +0000 UTC" firstStartedPulling="2025-10-07 22:20:00.395948655 +0000 UTC m=+5050.434529815" lastFinishedPulling="2025-10-07 22:20:06.42926181 +0000 UTC m=+5056.467842960" observedRunningTime="2025-10-07 22:20:07.511746374 +0000 UTC m=+5057.550327534" watchObservedRunningTime="2025-10-07 22:20:07.522791994 +0000 UTC m=+5057.561373144" Oct 07 22:20:09 crc kubenswrapper[4926]: I1007 22:20:09.087839 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:20:09 crc kubenswrapper[4926]: I1007 22:20:09.088162 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:20:10 crc kubenswrapper[4926]: I1007 22:20:10.142811 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wsmft" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="registry-server" probeResult="failure" output=< Oct 07 22:20:10 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Oct 07 22:20:10 crc kubenswrapper[4926]: > Oct 07 22:20:11 crc kubenswrapper[4926]: I1007 22:20:11.679301 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:20:11 crc kubenswrapper[4926]: E1007 22:20:11.679833 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:20:19 crc kubenswrapper[4926]: I1007 22:20:19.139645 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:20:19 crc kubenswrapper[4926]: I1007 22:20:19.188972 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:20:19 crc kubenswrapper[4926]: I1007 22:20:19.395288 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsmft"] Oct 07 22:20:20 crc kubenswrapper[4926]: I1007 22:20:20.648724 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wsmft" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="registry-server" containerID="cri-o://0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2" gracePeriod=2 Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.135590 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.279883 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-utilities\") pod \"a84b8ddc-1a7c-4e19-8320-6944446a3573\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.281426 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-utilities" (OuterVolumeSpecName: "utilities") pod "a84b8ddc-1a7c-4e19-8320-6944446a3573" (UID: "a84b8ddc-1a7c-4e19-8320-6944446a3573"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.281476 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-catalog-content\") pod \"a84b8ddc-1a7c-4e19-8320-6944446a3573\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.281520 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r4vg\" (UniqueName: \"kubernetes.io/projected/a84b8ddc-1a7c-4e19-8320-6944446a3573-kube-api-access-4r4vg\") pod \"a84b8ddc-1a7c-4e19-8320-6944446a3573\" (UID: \"a84b8ddc-1a7c-4e19-8320-6944446a3573\") " Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.282650 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.376959 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a84b8ddc-1a7c-4e19-8320-6944446a3573" (UID: "a84b8ddc-1a7c-4e19-8320-6944446a3573"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.385704 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a84b8ddc-1a7c-4e19-8320-6944446a3573-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.660982 4926 generic.go:334] "Generic (PLEG): container finished" podID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerID="0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2" exitCode=0 Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.661029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerDied","Data":"0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2"} Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.661083 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsmft" event={"ID":"a84b8ddc-1a7c-4e19-8320-6944446a3573","Type":"ContainerDied","Data":"1ff486d872468ae08c3d1dd1709001fd113ad2509249dfa114fec94d29e999b3"} Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.661106 4926 scope.go:117] "RemoveContainer" containerID="0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.661041 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsmft" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.708959 4926 scope.go:117] "RemoveContainer" containerID="6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.772330 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a84b8ddc-1a7c-4e19-8320-6944446a3573-kube-api-access-4r4vg" (OuterVolumeSpecName: "kube-api-access-4r4vg") pod "a84b8ddc-1a7c-4e19-8320-6944446a3573" (UID: "a84b8ddc-1a7c-4e19-8320-6944446a3573"). InnerVolumeSpecName "kube-api-access-4r4vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.784780 4926 scope.go:117] "RemoveContainer" containerID="5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.797491 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r4vg\" (UniqueName: \"kubernetes.io/projected/a84b8ddc-1a7c-4e19-8320-6944446a3573-kube-api-access-4r4vg\") on node \"crc\" DevicePath \"\"" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.922796 4926 scope.go:117] "RemoveContainer" containerID="0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2" Oct 07 22:20:21 crc kubenswrapper[4926]: E1007 22:20:21.924517 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2\": container with ID starting with 0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2 not found: ID does not exist" containerID="0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.924593 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2"} err="failed to get container status \"0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2\": rpc error: code = NotFound desc = could not find container \"0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2\": container with ID starting with 0e1097addcda0aaa5172ac1d7877e38f05c5591511628f453cf4d350c9c3c9e2 not found: ID does not exist" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.924636 4926 scope.go:117] "RemoveContainer" containerID="6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611" Oct 07 22:20:21 crc kubenswrapper[4926]: E1007 22:20:21.925166 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611\": container with ID starting with 6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611 not found: ID does not exist" containerID="6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.925229 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611"} err="failed to get container status \"6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611\": rpc error: code = NotFound desc = could not find container \"6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611\": container with ID starting with 6621085687ed011cc51c800c74eb82e32f0bcfb3cf13ea2607f4752602a81611 not found: ID does not exist" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.925260 4926 scope.go:117] "RemoveContainer" containerID="5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc" Oct 07 22:20:21 crc kubenswrapper[4926]: E1007 22:20:21.926038 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc\": container with ID starting with 5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc not found: ID does not exist" containerID="5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.926077 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc"} err="failed to get container status \"5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc\": rpc error: code = NotFound desc = could not find container \"5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc\": container with ID starting with 5c905e96bfd422ba8a2a6a6d873350e1925a9c4cb3d8f3e141f7268ebdc5bcdc not found: ID does not exist" Oct 07 22:20:21 crc kubenswrapper[4926]: I1007 22:20:21.992415 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsmft"] Oct 07 22:20:22 crc kubenswrapper[4926]: I1007 22:20:22.000358 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wsmft"] Oct 07 22:20:22 crc kubenswrapper[4926]: I1007 22:20:22.700648 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" path="/var/lib/kubelet/pods/a84b8ddc-1a7c-4e19-8320-6944446a3573/volumes" Oct 07 22:20:26 crc kubenswrapper[4926]: I1007 22:20:26.680893 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:20:26 crc kubenswrapper[4926]: E1007 22:20:26.681830 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:20:39 crc kubenswrapper[4926]: I1007 22:20:39.679294 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:20:39 crc kubenswrapper[4926]: E1007 22:20:39.680431 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:20:51 crc kubenswrapper[4926]: I1007 22:20:51.679798 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:20:51 crc kubenswrapper[4926]: E1007 22:20:51.680721 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:21:06 crc kubenswrapper[4926]: I1007 22:21:06.679457 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:21:06 crc kubenswrapper[4926]: E1007 22:21:06.680767 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:21:17 crc kubenswrapper[4926]: I1007 22:21:17.678580 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:21:17 crc kubenswrapper[4926]: E1007 22:21:17.679326 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:21:28 crc kubenswrapper[4926]: I1007 22:21:28.680011 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:21:28 crc kubenswrapper[4926]: E1007 22:21:28.681265 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:21:40 crc kubenswrapper[4926]: I1007 22:21:40.689316 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:21:40 crc kubenswrapper[4926]: E1007 22:21:40.690141 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:21:51 crc kubenswrapper[4926]: I1007 22:21:51.679832 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:21:51 crc kubenswrapper[4926]: E1007 22:21:51.680826 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:22:03 crc kubenswrapper[4926]: I1007 22:22:03.679620 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:22:03 crc kubenswrapper[4926]: E1007 22:22:03.680591 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:22:14 crc kubenswrapper[4926]: I1007 22:22:14.679550 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:22:14 crc kubenswrapper[4926]: E1007 22:22:14.680501 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:22:28 crc kubenswrapper[4926]: I1007 22:22:28.679561 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:22:28 crc kubenswrapper[4926]: E1007 22:22:28.680752 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:22:41 crc kubenswrapper[4926]: I1007 22:22:41.678725 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:22:41 crc kubenswrapper[4926]: E1007 22:22:41.679582 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:22:52 crc kubenswrapper[4926]: I1007 22:22:52.678928 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:22:52 crc kubenswrapper[4926]: E1007 22:22:52.679957 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.240079 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qdbr2"] Oct 07 22:23:00 crc kubenswrapper[4926]: E1007 22:23:00.240983 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="registry-server" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.240995 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="registry-server" Oct 07 22:23:00 crc kubenswrapper[4926]: E1007 22:23:00.241043 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="extract-utilities" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.241065 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="extract-utilities" Oct 07 22:23:00 crc kubenswrapper[4926]: E1007 22:23:00.241083 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="extract-content" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.241091 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="extract-content" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.241296 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a84b8ddc-1a7c-4e19-8320-6944446a3573" containerName="registry-server" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.242706 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.261347 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qdbr2"] Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.311720 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhcp5\" (UniqueName: \"kubernetes.io/projected/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-kube-api-access-qhcp5\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.311825 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-utilities\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.311893 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-catalog-content\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.414710 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-catalog-content\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.414939 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhcp5\" (UniqueName: \"kubernetes.io/projected/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-kube-api-access-qhcp5\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.415075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-utilities\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.415515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-catalog-content\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.416035 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-utilities\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.441682 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hzfbl"] Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.444175 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.456822 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhcp5\" (UniqueName: \"kubernetes.io/projected/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-kube-api-access-qhcp5\") pod \"certified-operators-qdbr2\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.476241 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzfbl"] Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.516764 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clh7b\" (UniqueName: \"kubernetes.io/projected/9964387c-32ee-4b5b-9279-f804aef27224-kube-api-access-clh7b\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.516999 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-utilities\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.517036 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-catalog-content\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.564852 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.618713 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-utilities\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.618782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-catalog-content\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.618848 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clh7b\" (UniqueName: \"kubernetes.io/projected/9964387c-32ee-4b5b-9279-f804aef27224-kube-api-access-clh7b\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.619833 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-utilities\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.620129 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-catalog-content\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.643187 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clh7b\" (UniqueName: \"kubernetes.io/projected/9964387c-32ee-4b5b-9279-f804aef27224-kube-api-access-clh7b\") pod \"community-operators-hzfbl\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:00 crc kubenswrapper[4926]: I1007 22:23:00.827941 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.090912 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qdbr2"] Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.332696 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzfbl"] Oct 07 22:23:01 crc kubenswrapper[4926]: W1007 22:23:01.335981 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9964387c_32ee_4b5b_9279_f804aef27224.slice/crio-8f85b35a888d69be699ec57a73e5df62b4d2cedd7bb88ae2c57a74c12e5dab0d WatchSource:0}: Error finding container 8f85b35a888d69be699ec57a73e5df62b4d2cedd7bb88ae2c57a74c12e5dab0d: Status 404 returned error can't find the container with id 8f85b35a888d69be699ec57a73e5df62b4d2cedd7bb88ae2c57a74c12e5dab0d Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.520727 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzfbl" event={"ID":"9964387c-32ee-4b5b-9279-f804aef27224","Type":"ContainerStarted","Data":"8f85b35a888d69be699ec57a73e5df62b4d2cedd7bb88ae2c57a74c12e5dab0d"} Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.526724 4926 generic.go:334] "Generic (PLEG): container finished" podID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerID="24a2f075658fd98abc69d415f5d05041a044c63f94db6a08f7fdf03e24061190" exitCode=0 Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.526771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerDied","Data":"24a2f075658fd98abc69d415f5d05041a044c63f94db6a08f7fdf03e24061190"} Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.526796 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerStarted","Data":"33d18275412148ea55ad17da39b1f611cf1dc3091b96cced9331bd8fdcd125a3"} Oct 07 22:23:01 crc kubenswrapper[4926]: I1007 22:23:01.528640 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:23:02 crc kubenswrapper[4926]: I1007 22:23:02.543163 4926 generic.go:334] "Generic (PLEG): container finished" podID="9964387c-32ee-4b5b-9279-f804aef27224" containerID="74bcce8af43384fee1f0e842b56b219c7e3810b159735c7ef47c136d85abc499" exitCode=0 Oct 07 22:23:02 crc kubenswrapper[4926]: I1007 22:23:02.543304 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzfbl" event={"ID":"9964387c-32ee-4b5b-9279-f804aef27224","Type":"ContainerDied","Data":"74bcce8af43384fee1f0e842b56b219c7e3810b159735c7ef47c136d85abc499"} Oct 07 22:23:03 crc kubenswrapper[4926]: I1007 22:23:03.560575 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerStarted","Data":"19f01cbe8af7cea4f81e29cfbbba7e474baf30643902fc54a148c79f51af9459"} Oct 07 22:23:04 crc kubenswrapper[4926]: I1007 22:23:04.576775 4926 generic.go:334] "Generic (PLEG): container finished" podID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerID="19f01cbe8af7cea4f81e29cfbbba7e474baf30643902fc54a148c79f51af9459" exitCode=0 Oct 07 22:23:04 crc kubenswrapper[4926]: I1007 22:23:04.576867 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerDied","Data":"19f01cbe8af7cea4f81e29cfbbba7e474baf30643902fc54a148c79f51af9459"} Oct 07 22:23:04 crc kubenswrapper[4926]: I1007 22:23:04.580115 4926 generic.go:334] "Generic (PLEG): container finished" podID="9964387c-32ee-4b5b-9279-f804aef27224" containerID="540b816c54e7b4c9d8bda7b6d37c8a5fe42ade1359c83ee2aee792ac3ddc7abf" exitCode=0 Oct 07 22:23:04 crc kubenswrapper[4926]: I1007 22:23:04.580179 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzfbl" event={"ID":"9964387c-32ee-4b5b-9279-f804aef27224","Type":"ContainerDied","Data":"540b816c54e7b4c9d8bda7b6d37c8a5fe42ade1359c83ee2aee792ac3ddc7abf"} Oct 07 22:23:06 crc kubenswrapper[4926]: I1007 22:23:06.611612 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerStarted","Data":"947c966fc5e9da982d1b869cd3fc80c8538c2aa56fe8410381c55f042b705235"} Oct 07 22:23:06 crc kubenswrapper[4926]: I1007 22:23:06.615986 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzfbl" event={"ID":"9964387c-32ee-4b5b-9279-f804aef27224","Type":"ContainerStarted","Data":"46832c24f4586f87f510b3ef2ba7ea4a08d88a2e43da79b96049cf460b308013"} Oct 07 22:23:06 crc kubenswrapper[4926]: I1007 22:23:06.634274 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qdbr2" podStartSLOduration=2.666393912 podStartE2EDuration="6.634253604s" podCreationTimestamp="2025-10-07 22:23:00 +0000 UTC" firstStartedPulling="2025-10-07 22:23:01.528388087 +0000 UTC m=+5231.566969237" lastFinishedPulling="2025-10-07 22:23:05.496247769 +0000 UTC m=+5235.534828929" observedRunningTime="2025-10-07 22:23:06.629320512 +0000 UTC m=+5236.667901682" watchObservedRunningTime="2025-10-07 22:23:06.634253604 +0000 UTC m=+5236.672834764" Oct 07 22:23:06 crc kubenswrapper[4926]: I1007 22:23:06.661745 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hzfbl" podStartSLOduration=3.561842238 podStartE2EDuration="6.661722137s" podCreationTimestamp="2025-10-07 22:23:00 +0000 UTC" firstStartedPulling="2025-10-07 22:23:02.545786652 +0000 UTC m=+5232.584367832" lastFinishedPulling="2025-10-07 22:23:05.645666581 +0000 UTC m=+5235.684247731" observedRunningTime="2025-10-07 22:23:06.651124201 +0000 UTC m=+5236.689705381" watchObservedRunningTime="2025-10-07 22:23:06.661722137 +0000 UTC m=+5236.700303297" Oct 07 22:23:07 crc kubenswrapper[4926]: I1007 22:23:07.679466 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:23:07 crc kubenswrapper[4926]: E1007 22:23:07.680145 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.565369 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.565738 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.657325 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.733891 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.828386 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.828461 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:10 crc kubenswrapper[4926]: I1007 22:23:10.880267 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:11 crc kubenswrapper[4926]: I1007 22:23:11.032913 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qdbr2"] Oct 07 22:23:11 crc kubenswrapper[4926]: I1007 22:23:11.758298 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:12 crc kubenswrapper[4926]: I1007 22:23:12.710142 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qdbr2" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="registry-server" containerID="cri-o://947c966fc5e9da982d1b869cd3fc80c8538c2aa56fe8410381c55f042b705235" gracePeriod=2 Oct 07 22:23:13 crc kubenswrapper[4926]: I1007 22:23:13.229408 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzfbl"] Oct 07 22:23:13 crc kubenswrapper[4926]: I1007 22:23:13.722088 4926 generic.go:334] "Generic (PLEG): container finished" podID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerID="947c966fc5e9da982d1b869cd3fc80c8538c2aa56fe8410381c55f042b705235" exitCode=0 Oct 07 22:23:13 crc kubenswrapper[4926]: I1007 22:23:13.722151 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerDied","Data":"947c966fc5e9da982d1b869cd3fc80c8538c2aa56fe8410381c55f042b705235"} Oct 07 22:23:13 crc kubenswrapper[4926]: I1007 22:23:13.722331 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hzfbl" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="registry-server" containerID="cri-o://46832c24f4586f87f510b3ef2ba7ea4a08d88a2e43da79b96049cf460b308013" gracePeriod=2 Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.147272 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.232049 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-utilities\") pod \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.232344 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhcp5\" (UniqueName: \"kubernetes.io/projected/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-kube-api-access-qhcp5\") pod \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.232433 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-catalog-content\") pod \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\" (UID: \"cf4540f6-5a03-4609-9345-e3e22dcf7f8b\") " Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.233515 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-utilities" (OuterVolumeSpecName: "utilities") pod "cf4540f6-5a03-4609-9345-e3e22dcf7f8b" (UID: "cf4540f6-5a03-4609-9345-e3e22dcf7f8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.244511 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-kube-api-access-qhcp5" (OuterVolumeSpecName: "kube-api-access-qhcp5") pod "cf4540f6-5a03-4609-9345-e3e22dcf7f8b" (UID: "cf4540f6-5a03-4609-9345-e3e22dcf7f8b"). InnerVolumeSpecName "kube-api-access-qhcp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.298223 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf4540f6-5a03-4609-9345-e3e22dcf7f8b" (UID: "cf4540f6-5a03-4609-9345-e3e22dcf7f8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.335595 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhcp5\" (UniqueName: \"kubernetes.io/projected/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-kube-api-access-qhcp5\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.335810 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.335872 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4540f6-5a03-4609-9345-e3e22dcf7f8b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.736656 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbr2" event={"ID":"cf4540f6-5a03-4609-9345-e3e22dcf7f8b","Type":"ContainerDied","Data":"33d18275412148ea55ad17da39b1f611cf1dc3091b96cced9331bd8fdcd125a3"} Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.736761 4926 scope.go:117] "RemoveContainer" containerID="947c966fc5e9da982d1b869cd3fc80c8538c2aa56fe8410381c55f042b705235" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.736880 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbr2" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.747873 4926 generic.go:334] "Generic (PLEG): container finished" podID="9964387c-32ee-4b5b-9279-f804aef27224" containerID="46832c24f4586f87f510b3ef2ba7ea4a08d88a2e43da79b96049cf460b308013" exitCode=0 Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.748024 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzfbl" event={"ID":"9964387c-32ee-4b5b-9279-f804aef27224","Type":"ContainerDied","Data":"46832c24f4586f87f510b3ef2ba7ea4a08d88a2e43da79b96049cf460b308013"} Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.760566 4926 scope.go:117] "RemoveContainer" containerID="19f01cbe8af7cea4f81e29cfbbba7e474baf30643902fc54a148c79f51af9459" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.820090 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qdbr2"] Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.830805 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qdbr2"] Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.852876 4926 scope.go:117] "RemoveContainer" containerID="24a2f075658fd98abc69d415f5d05041a044c63f94db6a08f7fdf03e24061190" Oct 07 22:23:14 crc kubenswrapper[4926]: I1007 22:23:14.968144 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.066581 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clh7b\" (UniqueName: \"kubernetes.io/projected/9964387c-32ee-4b5b-9279-f804aef27224-kube-api-access-clh7b\") pod \"9964387c-32ee-4b5b-9279-f804aef27224\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.066693 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-catalog-content\") pod \"9964387c-32ee-4b5b-9279-f804aef27224\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.066883 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-utilities\") pod \"9964387c-32ee-4b5b-9279-f804aef27224\" (UID: \"9964387c-32ee-4b5b-9279-f804aef27224\") " Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.067809 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-utilities" (OuterVolumeSpecName: "utilities") pod "9964387c-32ee-4b5b-9279-f804aef27224" (UID: "9964387c-32ee-4b5b-9279-f804aef27224"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.068120 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.134802 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9964387c-32ee-4b5b-9279-f804aef27224" (UID: "9964387c-32ee-4b5b-9279-f804aef27224"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.170283 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9964387c-32ee-4b5b-9279-f804aef27224-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.472502 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9964387c-32ee-4b5b-9279-f804aef27224-kube-api-access-clh7b" (OuterVolumeSpecName: "kube-api-access-clh7b") pod "9964387c-32ee-4b5b-9279-f804aef27224" (UID: "9964387c-32ee-4b5b-9279-f804aef27224"). InnerVolumeSpecName "kube-api-access-clh7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.477941 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clh7b\" (UniqueName: \"kubernetes.io/projected/9964387c-32ee-4b5b-9279-f804aef27224-kube-api-access-clh7b\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.764112 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzfbl" event={"ID":"9964387c-32ee-4b5b-9279-f804aef27224","Type":"ContainerDied","Data":"8f85b35a888d69be699ec57a73e5df62b4d2cedd7bb88ae2c57a74c12e5dab0d"} Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.764562 4926 scope.go:117] "RemoveContainer" containerID="46832c24f4586f87f510b3ef2ba7ea4a08d88a2e43da79b96049cf460b308013" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.764209 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzfbl" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.800873 4926 scope.go:117] "RemoveContainer" containerID="540b816c54e7b4c9d8bda7b6d37c8a5fe42ade1359c83ee2aee792ac3ddc7abf" Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.806497 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzfbl"] Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.821622 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hzfbl"] Oct 07 22:23:15 crc kubenswrapper[4926]: I1007 22:23:15.850839 4926 scope.go:117] "RemoveContainer" containerID="74bcce8af43384fee1f0e842b56b219c7e3810b159735c7ef47c136d85abc499" Oct 07 22:23:16 crc kubenswrapper[4926]: I1007 22:23:16.722783 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9964387c-32ee-4b5b-9279-f804aef27224" path="/var/lib/kubelet/pods/9964387c-32ee-4b5b-9279-f804aef27224/volumes" Oct 07 22:23:16 crc kubenswrapper[4926]: I1007 22:23:16.723809 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" path="/var/lib/kubelet/pods/cf4540f6-5a03-4609-9345-e3e22dcf7f8b/volumes" Oct 07 22:23:19 crc kubenswrapper[4926]: I1007 22:23:19.678938 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:23:19 crc kubenswrapper[4926]: E1007 22:23:19.680026 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:23:23 crc kubenswrapper[4926]: I1007 22:23:23.850640 4926 generic.go:334] "Generic (PLEG): container finished" podID="25a6326c-d6df-46b2-b646-0c37aa3bb0ed" containerID="01e8104ab7c378d2406efefd870eca413f48000e6da4ed5034105b83502af529" exitCode=0 Oct 07 22:23:23 crc kubenswrapper[4926]: I1007 22:23:23.850724 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"25a6326c-d6df-46b2-b646-0c37aa3bb0ed","Type":"ContainerDied","Data":"01e8104ab7c378d2406efefd870eca413f48000e6da4ed5034105b83502af529"} Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.234244 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.391661 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config-secret\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392116 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-temporary\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392305 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-config-data\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392375 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-workdir\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392436 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ssh-key\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn6h8\" (UniqueName: \"kubernetes.io/projected/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-kube-api-access-bn6h8\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392585 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392677 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392770 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ca-certs\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.392841 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config\") pod \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\" (UID: \"25a6326c-d6df-46b2-b646-0c37aa3bb0ed\") " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.393450 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-config-data" (OuterVolumeSpecName: "config-data") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.393693 4926 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.393719 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.397234 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.398751 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-kube-api-access-bn6h8" (OuterVolumeSpecName: "kube-api-access-bn6h8") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "kube-api-access-bn6h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.399139 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.421945 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.437649 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.451072 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.458860 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "25a6326c-d6df-46b2-b646-0c37aa3bb0ed" (UID: "25a6326c-d6df-46b2-b646-0c37aa3bb0ed"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495160 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495207 4926 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495221 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495233 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495245 4926 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495256 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.495266 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn6h8\" (UniqueName: \"kubernetes.io/projected/25a6326c-d6df-46b2-b646-0c37aa3bb0ed-kube-api-access-bn6h8\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.872986 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"25a6326c-d6df-46b2-b646-0c37aa3bb0ed","Type":"ContainerDied","Data":"2f5ce6af17884b917be1944f264e21cd76f0f4e35b1f8c0d9f7040733be60fec"} Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.873033 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f5ce6af17884b917be1944f264e21cd76f0f4e35b1f8c0d9f7040733be60fec" Oct 07 22:23:25 crc kubenswrapper[4926]: I1007 22:23:25.873092 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 07 22:23:26 crc kubenswrapper[4926]: I1007 22:23:26.008165 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 07 22:23:26 crc kubenswrapper[4926]: I1007 22:23:26.011702 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:23:32 crc kubenswrapper[4926]: I1007 22:23:32.679110 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:23:32 crc kubenswrapper[4926]: E1007 22:23:32.680980 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.452263 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.452913 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="extract-content" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.452933 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="extract-content" Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.452947 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="extract-content" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.452954 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="extract-content" Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.452976 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25a6326c-d6df-46b2-b646-0c37aa3bb0ed" containerName="tempest-tests-tempest-tests-runner" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.452984 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="25a6326c-d6df-46b2-b646-0c37aa3bb0ed" containerName="tempest-tests-tempest-tests-runner" Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.452997 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="extract-utilities" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453005 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="extract-utilities" Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.453018 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="registry-server" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453024 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="registry-server" Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.453036 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="registry-server" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453042 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="registry-server" Oct 07 22:23:34 crc kubenswrapper[4926]: E1007 22:23:34.453062 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="extract-utilities" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453069 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="extract-utilities" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453342 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf4540f6-5a03-4609-9345-e3e22dcf7f8b" containerName="registry-server" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453360 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="25a6326c-d6df-46b2-b646-0c37aa3bb0ed" containerName="tempest-tests-tempest-tests-runner" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.453371 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9964387c-32ee-4b5b-9279-f804aef27224" containerName="registry-server" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.454444 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.460518 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-86k4k" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.466890 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.615657 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.616655 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drfvs\" (UniqueName: \"kubernetes.io/projected/56b825d8-95d6-44ae-bad6-3f498d42f8db-kube-api-access-drfvs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.718942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.719165 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drfvs\" (UniqueName: \"kubernetes.io/projected/56b825d8-95d6-44ae-bad6-3f498d42f8db-kube-api-access-drfvs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.720025 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.747011 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drfvs\" (UniqueName: \"kubernetes.io/projected/56b825d8-95d6-44ae-bad6-3f498d42f8db-kube-api-access-drfvs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.762041 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"56b825d8-95d6-44ae-bad6-3f498d42f8db\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:34 crc kubenswrapper[4926]: I1007 22:23:34.782342 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 22:23:35 crc kubenswrapper[4926]: I1007 22:23:35.127373 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 07 22:23:36 crc kubenswrapper[4926]: I1007 22:23:36.020081 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"56b825d8-95d6-44ae-bad6-3f498d42f8db","Type":"ContainerStarted","Data":"4eee5ec7b8b7a4e41be41749eca71c8afe57af6461da7270deadd19362b81157"} Oct 07 22:23:38 crc kubenswrapper[4926]: I1007 22:23:38.049460 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"56b825d8-95d6-44ae-bad6-3f498d42f8db","Type":"ContainerStarted","Data":"9dfa46e26ccb272da377f3731bfbc03ef33285d2d030c682f9100ea319674a1a"} Oct 07 22:23:38 crc kubenswrapper[4926]: I1007 22:23:38.081357 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.141349979 podStartE2EDuration="4.08131661s" podCreationTimestamp="2025-10-07 22:23:34 +0000 UTC" firstStartedPulling="2025-10-07 22:23:35.121556005 +0000 UTC m=+5265.160137145" lastFinishedPulling="2025-10-07 22:23:37.061522616 +0000 UTC m=+5267.100103776" observedRunningTime="2025-10-07 22:23:38.067523492 +0000 UTC m=+5268.106104662" watchObservedRunningTime="2025-10-07 22:23:38.08131661 +0000 UTC m=+5268.119897770" Oct 07 22:23:47 crc kubenswrapper[4926]: I1007 22:23:47.679901 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:23:47 crc kubenswrapper[4926]: E1007 22:23:47.680839 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.869432 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-th6sx/must-gather-bn6kj"] Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.872641 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.880779 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-th6sx/must-gather-bn6kj"] Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.881104 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-th6sx"/"default-dockercfg-drnhx" Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.881332 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-th6sx"/"kube-root-ca.crt" Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.881615 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-th6sx"/"openshift-service-ca.crt" Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.934564 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5tks\" (UniqueName: \"kubernetes.io/projected/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-kube-api-access-d5tks\") pod \"must-gather-bn6kj\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:55 crc kubenswrapper[4926]: I1007 22:23:55.934641 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-must-gather-output\") pod \"must-gather-bn6kj\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:56 crc kubenswrapper[4926]: I1007 22:23:56.036214 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5tks\" (UniqueName: \"kubernetes.io/projected/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-kube-api-access-d5tks\") pod \"must-gather-bn6kj\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:56 crc kubenswrapper[4926]: I1007 22:23:56.036311 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-must-gather-output\") pod \"must-gather-bn6kj\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:56 crc kubenswrapper[4926]: I1007 22:23:56.036907 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-must-gather-output\") pod \"must-gather-bn6kj\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:56 crc kubenswrapper[4926]: I1007 22:23:56.055972 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5tks\" (UniqueName: \"kubernetes.io/projected/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-kube-api-access-d5tks\") pod \"must-gather-bn6kj\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:56 crc kubenswrapper[4926]: I1007 22:23:56.191864 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:23:56 crc kubenswrapper[4926]: W1007 22:23:56.750534 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a9a4b58_6d90_4dc4_a1ab_17d046c3adf6.slice/crio-898620ece972cda47cbfa0f4af561cc452074a0801f06a57d73aa5477222680a WatchSource:0}: Error finding container 898620ece972cda47cbfa0f4af561cc452074a0801f06a57d73aa5477222680a: Status 404 returned error can't find the container with id 898620ece972cda47cbfa0f4af561cc452074a0801f06a57d73aa5477222680a Oct 07 22:23:56 crc kubenswrapper[4926]: I1007 22:23:56.755862 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-th6sx/must-gather-bn6kj"] Oct 07 22:23:57 crc kubenswrapper[4926]: I1007 22:23:57.275598 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/must-gather-bn6kj" event={"ID":"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6","Type":"ContainerStarted","Data":"898620ece972cda47cbfa0f4af561cc452074a0801f06a57d73aa5477222680a"} Oct 07 22:24:01 crc kubenswrapper[4926]: I1007 22:24:01.679469 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:24:01 crc kubenswrapper[4926]: E1007 22:24:01.680205 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:24:04 crc kubenswrapper[4926]: I1007 22:24:04.362964 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/must-gather-bn6kj" event={"ID":"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6","Type":"ContainerStarted","Data":"9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232"} Oct 07 22:24:05 crc kubenswrapper[4926]: I1007 22:24:05.373933 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/must-gather-bn6kj" event={"ID":"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6","Type":"ContainerStarted","Data":"ce307b5cb97e5f4916afcb8025619cc9cdeccd1e6d79fe0eb61a9abda4b5b42b"} Oct 07 22:24:05 crc kubenswrapper[4926]: I1007 22:24:05.391649 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-th6sx/must-gather-bn6kj" podStartSLOduration=3.276133659 podStartE2EDuration="10.391633919s" podCreationTimestamp="2025-10-07 22:23:55 +0000 UTC" firstStartedPulling="2025-10-07 22:23:56.751916936 +0000 UTC m=+5286.790498096" lastFinishedPulling="2025-10-07 22:24:03.867417166 +0000 UTC m=+5293.905998356" observedRunningTime="2025-10-07 22:24:05.385730529 +0000 UTC m=+5295.424311679" watchObservedRunningTime="2025-10-07 22:24:05.391633919 +0000 UTC m=+5295.430215069" Oct 07 22:24:07 crc kubenswrapper[4926]: E1007 22:24:07.339010 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.58:44560->38.102.83.58:42335: write tcp 38.102.83.58:44560->38.102.83.58:42335: write: broken pipe Oct 07 22:24:08 crc kubenswrapper[4926]: I1007 22:24:08.857009 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-th6sx/crc-debug-jn88k"] Oct 07 22:24:08 crc kubenswrapper[4926]: I1007 22:24:08.858569 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:08 crc kubenswrapper[4926]: I1007 22:24:08.934683 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6tbs\" (UniqueName: \"kubernetes.io/projected/033372cb-d69e-43ff-8a26-192b957c0b51-kube-api-access-t6tbs\") pod \"crc-debug-jn88k\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:08 crc kubenswrapper[4926]: I1007 22:24:08.934905 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/033372cb-d69e-43ff-8a26-192b957c0b51-host\") pod \"crc-debug-jn88k\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:09 crc kubenswrapper[4926]: I1007 22:24:09.037265 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/033372cb-d69e-43ff-8a26-192b957c0b51-host\") pod \"crc-debug-jn88k\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:09 crc kubenswrapper[4926]: I1007 22:24:09.037412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/033372cb-d69e-43ff-8a26-192b957c0b51-host\") pod \"crc-debug-jn88k\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:09 crc kubenswrapper[4926]: I1007 22:24:09.037489 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6tbs\" (UniqueName: \"kubernetes.io/projected/033372cb-d69e-43ff-8a26-192b957c0b51-kube-api-access-t6tbs\") pod \"crc-debug-jn88k\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:09 crc kubenswrapper[4926]: I1007 22:24:09.056989 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6tbs\" (UniqueName: \"kubernetes.io/projected/033372cb-d69e-43ff-8a26-192b957c0b51-kube-api-access-t6tbs\") pod \"crc-debug-jn88k\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:09 crc kubenswrapper[4926]: I1007 22:24:09.189261 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:24:09 crc kubenswrapper[4926]: W1007 22:24:09.222215 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod033372cb_d69e_43ff_8a26_192b957c0b51.slice/crio-650f9093e464b3e5b4132fcce998db93b71c8029fe63222657e9e6ef9a474956 WatchSource:0}: Error finding container 650f9093e464b3e5b4132fcce998db93b71c8029fe63222657e9e6ef9a474956: Status 404 returned error can't find the container with id 650f9093e464b3e5b4132fcce998db93b71c8029fe63222657e9e6ef9a474956 Oct 07 22:24:09 crc kubenswrapper[4926]: I1007 22:24:09.417571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-jn88k" event={"ID":"033372cb-d69e-43ff-8a26-192b957c0b51","Type":"ContainerStarted","Data":"650f9093e464b3e5b4132fcce998db93b71c8029fe63222657e9e6ef9a474956"} Oct 07 22:24:13 crc kubenswrapper[4926]: I1007 22:24:13.679463 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:24:14 crc kubenswrapper[4926]: I1007 22:24:14.478681 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"cb5490e43946c352966fc9267e77332a505cb51138c598a5313fc5b04210ef95"} Oct 07 22:24:21 crc kubenswrapper[4926]: I1007 22:24:21.572416 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-jn88k" event={"ID":"033372cb-d69e-43ff-8a26-192b957c0b51","Type":"ContainerStarted","Data":"049b5c45d2ebbd2e734ee0dea11de605b4e27a5804f2c47adcc75c85c5749ee6"} Oct 07 22:24:21 crc kubenswrapper[4926]: I1007 22:24:21.591336 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-th6sx/crc-debug-jn88k" podStartSLOduration=2.2010535 podStartE2EDuration="13.591317919s" podCreationTimestamp="2025-10-07 22:24:08 +0000 UTC" firstStartedPulling="2025-10-07 22:24:09.224290248 +0000 UTC m=+5299.262871408" lastFinishedPulling="2025-10-07 22:24:20.614554667 +0000 UTC m=+5310.653135827" observedRunningTime="2025-10-07 22:24:21.589931669 +0000 UTC m=+5311.628512809" watchObservedRunningTime="2025-10-07 22:24:21.591317919 +0000 UTC m=+5311.629899089" Oct 07 22:25:36 crc kubenswrapper[4926]: I1007 22:25:36.678792 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6696d67754-xn5b4_d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c/barbican-api/0.log" Oct 07 22:25:36 crc kubenswrapper[4926]: I1007 22:25:36.741186 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6696d67754-xn5b4_d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c/barbican-api-log/0.log" Oct 07 22:25:36 crc kubenswrapper[4926]: I1007 22:25:36.939028 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6774c5f56-pjjjg_5badd158-2c74-425b-bc80-95ebba967905/barbican-keystone-listener/0.log" Oct 07 22:25:36 crc kubenswrapper[4926]: I1007 22:25:36.961149 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6774c5f56-pjjjg_5badd158-2c74-425b-bc80-95ebba967905/barbican-keystone-listener-log/0.log" Oct 07 22:25:37 crc kubenswrapper[4926]: I1007 22:25:37.148925 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-86dd45b86f-x4sv6_afa6dc20-30d9-48f1-886f-6b436a2f0678/barbican-worker/0.log" Oct 07 22:25:37 crc kubenswrapper[4926]: I1007 22:25:37.178322 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-86dd45b86f-x4sv6_afa6dc20-30d9-48f1-886f-6b436a2f0678/barbican-worker-log/0.log" Oct 07 22:25:37 crc kubenswrapper[4926]: I1007 22:25:37.404884 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn_9612131a-d029-4ed0-a6a0-f2562d6b34d2/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:37 crc kubenswrapper[4926]: I1007 22:25:37.636681 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/ceilometer-central-agent/0.log" Oct 07 22:25:37 crc kubenswrapper[4926]: I1007 22:25:37.701114 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/ceilometer-notification-agent/0.log" Oct 07 22:25:37 crc kubenswrapper[4926]: I1007 22:25:37.730571 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/proxy-httpd/0.log" Oct 07 22:25:38 crc kubenswrapper[4926]: I1007 22:25:38.495159 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/sg-core/0.log" Oct 07 22:25:38 crc kubenswrapper[4926]: I1007 22:25:38.657624 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_aabe891b-eda8-4359-b8c8-d4d99e309148/cinder-api/0.log" Oct 07 22:25:38 crc kubenswrapper[4926]: I1007 22:25:38.847276 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_aabe891b-eda8-4359-b8c8-d4d99e309148/cinder-api-log/0.log" Oct 07 22:25:38 crc kubenswrapper[4926]: I1007 22:25:38.945183 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_348d1b67-e2d4-4699-9890-3e3befa0133c/cinder-scheduler/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.108401 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_348d1b67-e2d4-4699-9890-3e3befa0133c/probe/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.209401 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm_67f58dbd-5fcb-47a6-a904-7e3c5232dd1e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.419386 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lhm27_3f3a65d1-ea19-4587-a483-ba006c9d4660/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.540881 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-rllwb_2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.639576 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c8d6588b7-27rms_f8f8c2fb-1675-4abc-8ae2-dbe68370e347/init/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.825594 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c8d6588b7-27rms_f8f8c2fb-1675-4abc-8ae2-dbe68370e347/init/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.895457 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq_ab42f719-a43c-425e-b0b3-aff4f812aefb/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:39 crc kubenswrapper[4926]: I1007 22:25:39.963181 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c8d6588b7-27rms_f8f8c2fb-1675-4abc-8ae2-dbe68370e347/dnsmasq-dns/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.102469 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_15c1a57a-dd1c-4955-ab28-86f0d09faffb/glance-httpd/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.169265 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_15c1a57a-dd1c-4955-ab28-86f0d09faffb/glance-log/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.284361 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e4b6a533-2f48-428f-b735-3ceac2e2d7c7/glance-httpd/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.336518 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e4b6a533-2f48-428f-b735-3ceac2e2d7c7/glance-log/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.409975 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-858b4c475d-fcp7q_c42be0d4-2ced-4d32-84f4-49389e2a0e82/horizon/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.636480 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf_52cd3a5e-b86a-49c9-84fd-387676280d79/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:40 crc kubenswrapper[4926]: I1007 22:25:40.909151 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-m9fn4_d64881a0-3a23-4075-94c1-46f52c09e5ef/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:41 crc kubenswrapper[4926]: I1007 22:25:41.007861 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-858b4c475d-fcp7q_c42be0d4-2ced-4d32-84f4-49389e2a0e82/horizon-log/0.log" Oct 07 22:25:41 crc kubenswrapper[4926]: I1007 22:25:41.146769 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29331241-qg8pm_ab4d73e0-30e0-4388-b7dc-28c97399930d/keystone-cron/0.log" Oct 07 22:25:41 crc kubenswrapper[4926]: I1007 22:25:41.329654 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b8c45bfd-abbf-432e-bb7f-70739b19c587/kube-state-metrics/0.log" Oct 07 22:25:41 crc kubenswrapper[4926]: I1007 22:25:41.429372 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-647f4cc4bb-zkt6b_434a0a19-516e-4685-be72-36976397fa42/keystone-api/0.log" Oct 07 22:25:41 crc kubenswrapper[4926]: I1007 22:25:41.493471 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr_4c8adb00-aa62-4d33-9b2f-a9b19b57e893/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:41 crc kubenswrapper[4926]: I1007 22:25:41.987129 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-b64c68dfc-k895v_c8be70c1-8342-414d-ba4b-098d40277bf6/neutron-api/0.log" Oct 07 22:25:42 crc kubenswrapper[4926]: I1007 22:25:42.008550 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-b64c68dfc-k895v_c8be70c1-8342-414d-ba4b-098d40277bf6/neutron-httpd/0.log" Oct 07 22:25:42 crc kubenswrapper[4926]: I1007 22:25:42.265712 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk_bd775971-aba6-499a-b63f-ee121ded14d4/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:43 crc kubenswrapper[4926]: I1007 22:25:43.020424 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b4601f7d-946c-4632-9778-44f70cafc757/nova-cell0-conductor-conductor/0.log" Oct 07 22:25:43 crc kubenswrapper[4926]: I1007 22:25:43.590697 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2f6780f7-788b-42cf-937e-750a222de004/nova-cell1-conductor-conductor/0.log" Oct 07 22:25:43 crc kubenswrapper[4926]: I1007 22:25:43.761540 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_08044ebc-ce53-4476-9884-83792098080d/nova-api-log/0.log" Oct 07 22:25:43 crc kubenswrapper[4926]: I1007 22:25:43.811080 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_08044ebc-ce53-4476-9884-83792098080d/nova-api-api/0.log" Oct 07 22:25:44 crc kubenswrapper[4926]: I1007 22:25:44.218945 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6e980975-58c0-4255-bc3e-df92773323cf/nova-cell1-novncproxy-novncproxy/0.log" Oct 07 22:25:44 crc kubenswrapper[4926]: I1007 22:25:44.250542 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-t7hkb_0273b882-c15f-4ca8-b786-60fb42c205ab/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:44 crc kubenswrapper[4926]: I1007 22:25:44.524425 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4496af78-6ebe-4aec-a04e-dcf4e8f446eb/nova-metadata-log/0.log" Oct 07 22:25:45 crc kubenswrapper[4926]: I1007 22:25:45.028642 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c5fe2fcd-3215-4278-90fc-3adc5cec14fe/nova-scheduler-scheduler/0.log" Oct 07 22:25:45 crc kubenswrapper[4926]: I1007 22:25:45.082956 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d169a79-a57d-458f-a790-3623f6dd0885/mysql-bootstrap/0.log" Oct 07 22:25:45 crc kubenswrapper[4926]: I1007 22:25:45.305023 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d169a79-a57d-458f-a790-3623f6dd0885/mysql-bootstrap/0.log" Oct 07 22:25:45 crc kubenswrapper[4926]: I1007 22:25:45.345715 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d169a79-a57d-458f-a790-3623f6dd0885/galera/0.log" Oct 07 22:25:45 crc kubenswrapper[4926]: I1007 22:25:45.584296 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7bf0bb92-3a4f-44ae-be03-e260cb246adb/mysql-bootstrap/0.log" Oct 07 22:25:46 crc kubenswrapper[4926]: I1007 22:25:46.190762 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7bf0bb92-3a4f-44ae-be03-e260cb246adb/mysql-bootstrap/0.log" Oct 07 22:25:46 crc kubenswrapper[4926]: I1007 22:25:46.310306 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7bf0bb92-3a4f-44ae-be03-e260cb246adb/galera/0.log" Oct 07 22:25:46 crc kubenswrapper[4926]: I1007 22:25:46.524944 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15/openstackclient/0.log" Oct 07 22:25:46 crc kubenswrapper[4926]: I1007 22:25:46.529429 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4496af78-6ebe-4aec-a04e-dcf4e8f446eb/nova-metadata-metadata/0.log" Oct 07 22:25:46 crc kubenswrapper[4926]: I1007 22:25:46.737264 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6qkqt_42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2/ovn-controller/0.log" Oct 07 22:25:46 crc kubenswrapper[4926]: I1007 22:25:46.912902 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5g8lv_5a9488ef-d520-4ff8-9eb2-8de1c72f03ce/openstack-network-exporter/0.log" Oct 07 22:25:47 crc kubenswrapper[4926]: I1007 22:25:47.031092 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovsdb-server-init/0.log" Oct 07 22:25:47 crc kubenswrapper[4926]: I1007 22:25:47.221340 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovsdb-server-init/0.log" Oct 07 22:25:47 crc kubenswrapper[4926]: I1007 22:25:47.940572 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovsdb-server/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.183544 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovs-vswitchd/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.215397 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cwx7z_bcecc1ad-1ce0-4b1a-8861-fc7d37276b12/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.398794 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee1a017c-a9f5-497d-9b4d-80f6f776be50/ovn-northd/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.432969 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee1a017c-a9f5-497d-9b4d-80f6f776be50/openstack-network-exporter/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.683933 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b2704dcf-cd03-4f91-af4e-ea708f3c0534/openstack-network-exporter/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.688400 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b2704dcf-cd03-4f91-af4e-ea708f3c0534/ovsdbserver-nb/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.895525 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fdc18634-337d-4562-a6e5-a20df1f53028/openstack-network-exporter/0.log" Oct 07 22:25:48 crc kubenswrapper[4926]: I1007 22:25:48.927069 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fdc18634-337d-4562-a6e5-a20df1f53028/ovsdbserver-sb/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.305573 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-849bb7d7cd-xqmg2_1af8a1d9-b36b-4f59-905c-12d2245ae3cf/placement-api/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.340616 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-849bb7d7cd-xqmg2_1af8a1d9-b36b-4f59-905c-12d2245ae3cf/placement-log/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.500148 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/init-config-reloader/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.685477 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/prometheus/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.686955 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/config-reloader/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.717017 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/init-config-reloader/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.845506 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/thanos-sidecar/0.log" Oct 07 22:25:49 crc kubenswrapper[4926]: I1007 22:25:49.947937 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d5d417c9-ac1d-4b58-9506-bc6dd17a71b1/setup-container/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.151109 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d5d417c9-ac1d-4b58-9506-bc6dd17a71b1/setup-container/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.207480 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d5d417c9-ac1d-4b58-9506-bc6dd17a71b1/rabbitmq/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.381502 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_5b0d7388-05f4-4df6-b093-c9157a4658b0/setup-container/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.436558 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_038e2e94-c12a-4d54-8b48-96d08012ab97/memcached/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.542279 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_5b0d7388-05f4-4df6-b093-c9157a4658b0/setup-container/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.548740 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_5b0d7388-05f4-4df6-b093-c9157a4658b0/rabbitmq/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.648780 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9052310e-1ddc-4b9b-a3bc-ced3dc97ab21/setup-container/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.806803 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9052310e-1ddc-4b9b-a3bc-ced3dc97ab21/setup-container/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.842798 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd_28a80874-caa4-48eb-9cbc-999e2db615fb/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:50 crc kubenswrapper[4926]: I1007 22:25:50.889223 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9052310e-1ddc-4b9b-a3bc-ced3dc97ab21/rabbitmq/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.058435 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-hnftc_b581b351-be97-4417-9c19-58e962bf821e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.091308 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc_b9c1704c-58c8-407c-bafc-c23a19cf900b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.250512 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kbwzm_67068a37-7d4a-47ca-9971-89e1e7a49a25/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.354126 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-5rwsd_69fe71d5-112a-45aa-a043-8629073ec0c2/ssh-known-hosts-edpm-deployment/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.572864 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55b995649c-qwlp9_28011eca-ae6f-42da-9e6b-27289eb19b2e/proxy-server/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.641332 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55b995649c-qwlp9_28011eca-ae6f-42da-9e6b-27289eb19b2e/proxy-httpd/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.688286 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-fz74j_f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f/swift-ring-rebalance/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.759698 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-auditor/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.837756 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-reaper/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.910109 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-replicator/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.941371 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-auditor/0.log" Oct 07 22:25:51 crc kubenswrapper[4926]: I1007 22:25:51.988023 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-server/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.077297 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-server/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.081481 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-replicator/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.147949 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-updater/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.213738 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-auditor/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.293127 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-expirer/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.294342 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-replicator/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.367262 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-server/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.443419 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-updater/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.468166 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/swift-recon-cron/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.468876 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/rsync/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.691563 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v99xb_013ef09a-4927-4ca6-9beb-0537614a9187/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.722132 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_25a6326c-d6df-46b2-b646-0c37aa3bb0ed/tempest-tests-tempest-tests-runner/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.858502 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_56b825d8-95d6-44ae-bad6-3f498d42f8db/test-operator-logs-container/0.log" Oct 07 22:25:52 crc kubenswrapper[4926]: I1007 22:25:52.976634 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx_a94c25aa-d7cb-4329-8dbc-d52c8b511cf3/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:25:53 crc kubenswrapper[4926]: I1007 22:25:53.876701 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_7d2feaba-0aff-4563-94f6-939db1594277/watcher-applier/0.log" Oct 07 22:25:54 crc kubenswrapper[4926]: I1007 22:25:54.202740 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_3d69b703-1e74-4e92-943d-6c73f029020e/watcher-api-log/0.log" Oct 07 22:25:56 crc kubenswrapper[4926]: I1007 22:25:56.571153 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_d9756c29-8c2e-49c9-8bf2-f9a51710d7d7/watcher-decision-engine/0.log" Oct 07 22:25:57 crc kubenswrapper[4926]: I1007 22:25:57.364930 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_3d69b703-1e74-4e92-943d-6c73f029020e/watcher-api/0.log" Oct 07 22:26:33 crc kubenswrapper[4926]: I1007 22:26:33.209364 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:26:33 crc kubenswrapper[4926]: I1007 22:26:33.210301 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:26:36 crc kubenswrapper[4926]: I1007 22:26:36.953039 4926 generic.go:334] "Generic (PLEG): container finished" podID="033372cb-d69e-43ff-8a26-192b957c0b51" containerID="049b5c45d2ebbd2e734ee0dea11de605b4e27a5804f2c47adcc75c85c5749ee6" exitCode=0 Oct 07 22:26:36 crc kubenswrapper[4926]: I1007 22:26:36.953124 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-jn88k" event={"ID":"033372cb-d69e-43ff-8a26-192b957c0b51","Type":"ContainerDied","Data":"049b5c45d2ebbd2e734ee0dea11de605b4e27a5804f2c47adcc75c85c5749ee6"} Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.107883 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.148386 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-th6sx/crc-debug-jn88k"] Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.159942 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-th6sx/crc-debug-jn88k"] Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.181983 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/033372cb-d69e-43ff-8a26-192b957c0b51-host\") pod \"033372cb-d69e-43ff-8a26-192b957c0b51\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.182102 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/033372cb-d69e-43ff-8a26-192b957c0b51-host" (OuterVolumeSpecName: "host") pod "033372cb-d69e-43ff-8a26-192b957c0b51" (UID: "033372cb-d69e-43ff-8a26-192b957c0b51"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.182217 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6tbs\" (UniqueName: \"kubernetes.io/projected/033372cb-d69e-43ff-8a26-192b957c0b51-kube-api-access-t6tbs\") pod \"033372cb-d69e-43ff-8a26-192b957c0b51\" (UID: \"033372cb-d69e-43ff-8a26-192b957c0b51\") " Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.182759 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/033372cb-d69e-43ff-8a26-192b957c0b51-host\") on node \"crc\" DevicePath \"\"" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.188371 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/033372cb-d69e-43ff-8a26-192b957c0b51-kube-api-access-t6tbs" (OuterVolumeSpecName: "kube-api-access-t6tbs") pod "033372cb-d69e-43ff-8a26-192b957c0b51" (UID: "033372cb-d69e-43ff-8a26-192b957c0b51"). InnerVolumeSpecName "kube-api-access-t6tbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.284322 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6tbs\" (UniqueName: \"kubernetes.io/projected/033372cb-d69e-43ff-8a26-192b957c0b51-kube-api-access-t6tbs\") on node \"crc\" DevicePath \"\"" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.700360 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="033372cb-d69e-43ff-8a26-192b957c0b51" path="/var/lib/kubelet/pods/033372cb-d69e-43ff-8a26-192b957c0b51/volumes" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.979893 4926 scope.go:117] "RemoveContainer" containerID="049b5c45d2ebbd2e734ee0dea11de605b4e27a5804f2c47adcc75c85c5749ee6" Oct 07 22:26:38 crc kubenswrapper[4926]: I1007 22:26:38.979940 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-jn88k" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.334139 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-th6sx/crc-debug-9w47h"] Oct 07 22:26:39 crc kubenswrapper[4926]: E1007 22:26:39.334780 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="033372cb-d69e-43ff-8a26-192b957c0b51" containerName="container-00" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.334802 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="033372cb-d69e-43ff-8a26-192b957c0b51" containerName="container-00" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.335191 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="033372cb-d69e-43ff-8a26-192b957c0b51" containerName="container-00" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.336333 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.415882 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fww2x\" (UniqueName: \"kubernetes.io/projected/5870a3dc-2ac4-4be1-9e69-551bca40efff-kube-api-access-fww2x\") pod \"crc-debug-9w47h\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.415997 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5870a3dc-2ac4-4be1-9e69-551bca40efff-host\") pod \"crc-debug-9w47h\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.517933 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fww2x\" (UniqueName: \"kubernetes.io/projected/5870a3dc-2ac4-4be1-9e69-551bca40efff-kube-api-access-fww2x\") pod \"crc-debug-9w47h\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.518101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5870a3dc-2ac4-4be1-9e69-551bca40efff-host\") pod \"crc-debug-9w47h\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.518317 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5870a3dc-2ac4-4be1-9e69-551bca40efff-host\") pod \"crc-debug-9w47h\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.551922 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fww2x\" (UniqueName: \"kubernetes.io/projected/5870a3dc-2ac4-4be1-9e69-551bca40efff-kube-api-access-fww2x\") pod \"crc-debug-9w47h\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.665232 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:39 crc kubenswrapper[4926]: I1007 22:26:39.994594 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-9w47h" event={"ID":"5870a3dc-2ac4-4be1-9e69-551bca40efff","Type":"ContainerStarted","Data":"39b40a15764a0a4360574020921e12dc06bd62e2e8d6de653ed34c3eff22b89d"} Oct 07 22:26:41 crc kubenswrapper[4926]: I1007 22:26:41.011527 4926 generic.go:334] "Generic (PLEG): container finished" podID="5870a3dc-2ac4-4be1-9e69-551bca40efff" containerID="605ab15b8f73526a347a3fc976ae82287f612974090f32b79f9b144552d0f1b3" exitCode=0 Oct 07 22:26:41 crc kubenswrapper[4926]: I1007 22:26:41.011615 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-9w47h" event={"ID":"5870a3dc-2ac4-4be1-9e69-551bca40efff","Type":"ContainerDied","Data":"605ab15b8f73526a347a3fc976ae82287f612974090f32b79f9b144552d0f1b3"} Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.145428 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.275244 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fww2x\" (UniqueName: \"kubernetes.io/projected/5870a3dc-2ac4-4be1-9e69-551bca40efff-kube-api-access-fww2x\") pod \"5870a3dc-2ac4-4be1-9e69-551bca40efff\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.275409 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5870a3dc-2ac4-4be1-9e69-551bca40efff-host\") pod \"5870a3dc-2ac4-4be1-9e69-551bca40efff\" (UID: \"5870a3dc-2ac4-4be1-9e69-551bca40efff\") " Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.275462 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5870a3dc-2ac4-4be1-9e69-551bca40efff-host" (OuterVolumeSpecName: "host") pod "5870a3dc-2ac4-4be1-9e69-551bca40efff" (UID: "5870a3dc-2ac4-4be1-9e69-551bca40efff"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.276169 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5870a3dc-2ac4-4be1-9e69-551bca40efff-host\") on node \"crc\" DevicePath \"\"" Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.281908 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5870a3dc-2ac4-4be1-9e69-551bca40efff-kube-api-access-fww2x" (OuterVolumeSpecName: "kube-api-access-fww2x") pod "5870a3dc-2ac4-4be1-9e69-551bca40efff" (UID: "5870a3dc-2ac4-4be1-9e69-551bca40efff"). InnerVolumeSpecName "kube-api-access-fww2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:26:42 crc kubenswrapper[4926]: I1007 22:26:42.377483 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fww2x\" (UniqueName: \"kubernetes.io/projected/5870a3dc-2ac4-4be1-9e69-551bca40efff-kube-api-access-fww2x\") on node \"crc\" DevicePath \"\"" Oct 07 22:26:43 crc kubenswrapper[4926]: I1007 22:26:43.031401 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-9w47h" event={"ID":"5870a3dc-2ac4-4be1-9e69-551bca40efff","Type":"ContainerDied","Data":"39b40a15764a0a4360574020921e12dc06bd62e2e8d6de653ed34c3eff22b89d"} Oct 07 22:26:43 crc kubenswrapper[4926]: I1007 22:26:43.031684 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39b40a15764a0a4360574020921e12dc06bd62e2e8d6de653ed34c3eff22b89d" Oct 07 22:26:43 crc kubenswrapper[4926]: I1007 22:26:43.031489 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-9w47h" Oct 07 22:26:49 crc kubenswrapper[4926]: I1007 22:26:49.548733 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-th6sx/crc-debug-9w47h"] Oct 07 22:26:49 crc kubenswrapper[4926]: I1007 22:26:49.556290 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-th6sx/crc-debug-9w47h"] Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.705655 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5870a3dc-2ac4-4be1-9e69-551bca40efff" path="/var/lib/kubelet/pods/5870a3dc-2ac4-4be1-9e69-551bca40efff/volumes" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.760870 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-th6sx/crc-debug-94tmz"] Oct 07 22:26:50 crc kubenswrapper[4926]: E1007 22:26:50.761338 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5870a3dc-2ac4-4be1-9e69-551bca40efff" containerName="container-00" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.761359 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5870a3dc-2ac4-4be1-9e69-551bca40efff" containerName="container-00" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.761594 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5870a3dc-2ac4-4be1-9e69-551bca40efff" containerName="container-00" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.762381 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.808330 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-host\") pod \"crc-debug-94tmz\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.808956 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqt4c\" (UniqueName: \"kubernetes.io/projected/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-kube-api-access-lqt4c\") pod \"crc-debug-94tmz\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.911420 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-host\") pod \"crc-debug-94tmz\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.911578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-host\") pod \"crc-debug-94tmz\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.911623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqt4c\" (UniqueName: \"kubernetes.io/projected/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-kube-api-access-lqt4c\") pod \"crc-debug-94tmz\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:50 crc kubenswrapper[4926]: I1007 22:26:50.933655 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqt4c\" (UniqueName: \"kubernetes.io/projected/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-kube-api-access-lqt4c\") pod \"crc-debug-94tmz\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:51 crc kubenswrapper[4926]: I1007 22:26:51.089456 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:52 crc kubenswrapper[4926]: I1007 22:26:52.111956 4926 generic.go:334] "Generic (PLEG): container finished" podID="60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" containerID="0e7f849123a173f898ec7977d40bfeba83f34cfcdc64b870de10faa4fa6d3c5f" exitCode=0 Oct 07 22:26:52 crc kubenswrapper[4926]: I1007 22:26:52.112061 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-94tmz" event={"ID":"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0","Type":"ContainerDied","Data":"0e7f849123a173f898ec7977d40bfeba83f34cfcdc64b870de10faa4fa6d3c5f"} Oct 07 22:26:52 crc kubenswrapper[4926]: I1007 22:26:52.112341 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/crc-debug-94tmz" event={"ID":"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0","Type":"ContainerStarted","Data":"578ba5cd1e046a082171ca86782ea72a4c11196bcdc45a53063771df29dc3be1"} Oct 07 22:26:52 crc kubenswrapper[4926]: I1007 22:26:52.188423 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-th6sx/crc-debug-94tmz"] Oct 07 22:26:52 crc kubenswrapper[4926]: I1007 22:26:52.202302 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-th6sx/crc-debug-94tmz"] Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.235261 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.263594 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-host\") pod \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.263754 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-host" (OuterVolumeSpecName: "host") pod "60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" (UID: "60bbdff6-88d5-4053-bf6a-ab1ae750f2a0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.264120 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqt4c\" (UniqueName: \"kubernetes.io/projected/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-kube-api-access-lqt4c\") pod \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\" (UID: \"60bbdff6-88d5-4053-bf6a-ab1ae750f2a0\") " Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.265002 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-host\") on node \"crc\" DevicePath \"\"" Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.270743 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-kube-api-access-lqt4c" (OuterVolumeSpecName: "kube-api-access-lqt4c") pod "60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" (UID: "60bbdff6-88d5-4053-bf6a-ab1ae750f2a0"). InnerVolumeSpecName "kube-api-access-lqt4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.366611 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqt4c\" (UniqueName: \"kubernetes.io/projected/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0-kube-api-access-lqt4c\") on node \"crc\" DevicePath \"\"" Oct 07 22:26:53 crc kubenswrapper[4926]: I1007 22:26:53.862247 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/util/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.129053 4926 scope.go:117] "RemoveContainer" containerID="0e7f849123a173f898ec7977d40bfeba83f34cfcdc64b870de10faa4fa6d3c5f" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.129178 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/crc-debug-94tmz" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.211392 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/util/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.250177 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/pull/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.284657 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/pull/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.421313 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/pull/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.427330 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/extract/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.444242 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/util/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.620823 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-7ltns_fea578e9-bbed-446d-beb9-5fcab1884b60/kube-rbac-proxy/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.672606 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-bljj9_313c922a-ffe2-4d92-a221-bde117d0f196/kube-rbac-proxy/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.687630 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-7ltns_fea578e9-bbed-446d-beb9-5fcab1884b60/manager/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.690608 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" path="/var/lib/kubelet/pods/60bbdff6-88d5-4053-bf6a-ab1ae750f2a0/volumes" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.834053 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-bljj9_313c922a-ffe2-4d92-a221-bde117d0f196/manager/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.854539 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-pwsh8_a2ab6887-9ac0-4beb-8da0-8cb27da2ff66/kube-rbac-proxy/0.log" Oct 07 22:26:54 crc kubenswrapper[4926]: I1007 22:26:54.913012 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-pwsh8_a2ab6887-9ac0-4beb-8da0-8cb27da2ff66/manager/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.025731 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-jnbmb_bd230ff4-3566-43dc-be73-46cfe038bd76/kube-rbac-proxy/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.160259 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-jnbmb_bd230ff4-3566-43dc-be73-46cfe038bd76/manager/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.212368 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-mwlfv_fc46d600-5c99-4ce4-9beb-06376dbacd8a/kube-rbac-proxy/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.246057 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-mwlfv_fc46d600-5c99-4ce4-9beb-06376dbacd8a/manager/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.343252 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-nm2mk_6cb125b4-a707-4506-80fc-76bc981f1560/kube-rbac-proxy/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.416482 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-nm2mk_6cb125b4-a707-4506-80fc-76bc981f1560/manager/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.488183 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-cgkzf_5517a3d5-4f45-44db-9552-fcd8de7356bf/kube-rbac-proxy/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.647598 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-zz2jk_d1f7536d-1b34-4d39-9249-b9959aa2a00f/kube-rbac-proxy/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.690386 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-cgkzf_5517a3d5-4f45-44db-9552-fcd8de7356bf/manager/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.730375 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-zz2jk_d1f7536d-1b34-4d39-9249-b9959aa2a00f/manager/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.838907 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-wxrzg_cefed49a-2a06-49a6-917e-ff05d6457bca/kube-rbac-proxy/0.log" Oct 07 22:26:55 crc kubenswrapper[4926]: I1007 22:26:55.931158 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-wxrzg_cefed49a-2a06-49a6-917e-ff05d6457bca/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.034721 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-tmh94_bd7d6078-1c87-459e-b83e-c408537f82a1/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.078801 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-tmh94_bd7d6078-1c87-459e-b83e-c408537f82a1/kube-rbac-proxy/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.188570 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-7spjf_09117b6f-d508-460f-8848-ad0ba4f7c1be/kube-rbac-proxy/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.266000 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-7spjf_09117b6f-d508-460f-8848-ad0ba4f7c1be/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.371836 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-tjfcp_ce47a525-41f3-4beb-b9b3-28d85831fdda/kube-rbac-proxy/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.453775 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-tjfcp_ce47a525-41f3-4beb-b9b3-28d85831fdda/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.503833 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4xz96_0d0347d8-c6bc-4bd5-b75a-28ba94906a6d/kube-rbac-proxy/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.615636 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4xz96_0d0347d8-c6bc-4bd5-b75a-28ba94906a6d/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.650944 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-rxdc8_2d16f135-1997-4bb6-9cb3-dae655640a01/kube-rbac-proxy/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.708239 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-rxdc8_2d16f135-1997-4bb6-9cb3-dae655640a01/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.850799 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc_11159753-a963-44c9-bb63-53340a200de7/manager/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.888516 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc_11159753-a963-44c9-bb63-53340a200de7/kube-rbac-proxy/0.log" Oct 07 22:26:56 crc kubenswrapper[4926]: I1007 22:26:56.987791 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-9554d4fbd-2cvhj_78f2ab6d-0297-49e4-91ff-ae7c8251fb0e/kube-rbac-proxy/0.log" Oct 07 22:26:57 crc kubenswrapper[4926]: I1007 22:26:57.138786 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8578696b-282fd_37e8795d-6050-4b61-9282-f3af17a926c8/kube-rbac-proxy/0.log" Oct 07 22:26:57 crc kubenswrapper[4926]: I1007 22:26:57.357303 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8578696b-282fd_37e8795d-6050-4b61-9282-f3af17a926c8/operator/0.log" Oct 07 22:26:57 crc kubenswrapper[4926]: I1007 22:26:57.397571 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hbjzj_240049fd-6fcf-493a-933e-c64ebca105b3/registry-server/0.log" Oct 07 22:26:57 crc kubenswrapper[4926]: I1007 22:26:57.409927 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-tsklh_f490b56a-2c4e-49ea-93fc-f2d8fb9454fd/kube-rbac-proxy/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.187695 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-47m5t_d039dde6-16e1-42dd-a19f-b564a3939837/kube-rbac-proxy/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.247051 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-tsklh_f490b56a-2c4e-49ea-93fc-f2d8fb9454fd/manager/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.262701 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-47m5t_d039dde6-16e1-42dd-a19f-b564a3939837/manager/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.262886 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-9554d4fbd-2cvhj_78f2ab6d-0297-49e4-91ff-ae7c8251fb0e/manager/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.383719 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5_fb36bed3-c7f6-43ad-b250-da42097b673e/operator/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.490613 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-59xh2_35b344ad-2ef7-48e3-8643-e93e974fb326/manager/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.518264 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-59xh2_35b344ad-2ef7-48e3-8643-e93e974fb326/kube-rbac-proxy/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.615725 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-w2bgf_659a6e11-31f5-4a95-9eb1-3f61d8891ace/kube-rbac-proxy/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.679855 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-rtwpd_f825a6cc-9414-44f5-98a2-df2c3f1e825c/kube-rbac-proxy/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.850684 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-rtwpd_f825a6cc-9414-44f5-98a2-df2c3f1e825c/manager/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.864037 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-w2bgf_659a6e11-31f5-4a95-9eb1-3f61d8891ace/manager/0.log" Oct 07 22:26:58 crc kubenswrapper[4926]: I1007 22:26:58.925756 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-77fdd5cc9b-q8mkp_45317f9d-c1aa-4f42-b52f-f910a7082d1c/kube-rbac-proxy/0.log" Oct 07 22:26:59 crc kubenswrapper[4926]: I1007 22:26:59.070404 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-77fdd5cc9b-q8mkp_45317f9d-c1aa-4f42-b52f-f910a7082d1c/manager/0.log" Oct 07 22:27:03 crc kubenswrapper[4926]: I1007 22:27:03.209059 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:27:03 crc kubenswrapper[4926]: I1007 22:27:03.209537 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:27:17 crc kubenswrapper[4926]: I1007 22:27:17.214974 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6mlnn_5026deaa-f35e-4c06-9c90-cff95d6599a2/control-plane-machine-set-operator/0.log" Oct 07 22:27:17 crc kubenswrapper[4926]: I1007 22:27:17.321429 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qzh7t_24f2e598-e97c-4583-bc67-cd8afcf6065e/kube-rbac-proxy/0.log" Oct 07 22:27:17 crc kubenswrapper[4926]: I1007 22:27:17.321873 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qzh7t_24f2e598-e97c-4583-bc67-cd8afcf6065e/machine-api-operator/0.log" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.086458 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t4mpq"] Oct 07 22:27:27 crc kubenswrapper[4926]: E1007 22:27:27.087589 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" containerName="container-00" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.087607 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" containerName="container-00" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.087851 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bbdff6-88d5-4053-bf6a-ab1ae750f2a0" containerName="container-00" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.089404 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.118001 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4mpq"] Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.196988 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-catalog-content\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.197443 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87m95\" (UniqueName: \"kubernetes.io/projected/03a5135a-53b3-4594-adcb-44de3c6918f9-kube-api-access-87m95\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.197533 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-utilities\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.298977 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-utilities\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.299473 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-catalog-content\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.299599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-utilities\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.299771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87m95\" (UniqueName: \"kubernetes.io/projected/03a5135a-53b3-4594-adcb-44de3c6918f9-kube-api-access-87m95\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.300276 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-catalog-content\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.331228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87m95\" (UniqueName: \"kubernetes.io/projected/03a5135a-53b3-4594-adcb-44de3c6918f9-kube-api-access-87m95\") pod \"redhat-marketplace-t4mpq\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.417998 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:27 crc kubenswrapper[4926]: I1007 22:27:27.958079 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4mpq"] Oct 07 22:27:28 crc kubenswrapper[4926]: I1007 22:27:28.485964 4926 generic.go:334] "Generic (PLEG): container finished" podID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerID="89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14" exitCode=0 Oct 07 22:27:28 crc kubenswrapper[4926]: I1007 22:27:28.486054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerDied","Data":"89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14"} Oct 07 22:27:28 crc kubenswrapper[4926]: I1007 22:27:28.486250 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerStarted","Data":"48d612f40a392e278186ff8a34456d2a0a44317f0d15123ccc7850b41bc68bb2"} Oct 07 22:27:30 crc kubenswrapper[4926]: I1007 22:27:30.511919 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerStarted","Data":"c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a"} Oct 07 22:27:31 crc kubenswrapper[4926]: I1007 22:27:31.522945 4926 generic.go:334] "Generic (PLEG): container finished" podID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerID="c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a" exitCode=0 Oct 07 22:27:31 crc kubenswrapper[4926]: I1007 22:27:31.523062 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerDied","Data":"c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a"} Oct 07 22:27:32 crc kubenswrapper[4926]: I1007 22:27:32.208117 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-v56pb_742ac735-fc39-4ab6-a072-834db359b8d8/cert-manager-controller/0.log" Oct 07 22:27:32 crc kubenswrapper[4926]: I1007 22:27:32.372835 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-77j9j_3aa5ce8c-a725-4bd2-a504-887c768a64b6/cert-manager-cainjector/0.log" Oct 07 22:27:32 crc kubenswrapper[4926]: I1007 22:27:32.461777 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dd6ts_07446e4f-68a2-422e-b303-e3637d956a4c/cert-manager-webhook/0.log" Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.209903 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.212600 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.212709 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.214107 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb5490e43946c352966fc9267e77332a505cb51138c598a5313fc5b04210ef95"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.214287 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://cb5490e43946c352966fc9267e77332a505cb51138c598a5313fc5b04210ef95" gracePeriod=600 Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.611769 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="cb5490e43946c352966fc9267e77332a505cb51138c598a5313fc5b04210ef95" exitCode=0 Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.612274 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"cb5490e43946c352966fc9267e77332a505cb51138c598a5313fc5b04210ef95"} Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.612310 4926 scope.go:117] "RemoveContainer" containerID="0b783d1162b59176887330e7d4f305f487da6fa84082b957275523ebcc16eb58" Oct 07 22:27:33 crc kubenswrapper[4926]: I1007 22:27:33.633623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerStarted","Data":"29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73"} Oct 07 22:27:34 crc kubenswrapper[4926]: I1007 22:27:34.645744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984"} Oct 07 22:27:37 crc kubenswrapper[4926]: I1007 22:27:37.418397 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:37 crc kubenswrapper[4926]: I1007 22:27:37.419144 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:37 crc kubenswrapper[4926]: I1007 22:27:37.495057 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:37 crc kubenswrapper[4926]: I1007 22:27:37.511971 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t4mpq" podStartSLOduration=6.195253495 podStartE2EDuration="10.511956885s" podCreationTimestamp="2025-10-07 22:27:27 +0000 UTC" firstStartedPulling="2025-10-07 22:27:28.488417195 +0000 UTC m=+5498.526998345" lastFinishedPulling="2025-10-07 22:27:32.805120575 +0000 UTC m=+5502.843701735" observedRunningTime="2025-10-07 22:27:33.674079555 +0000 UTC m=+5503.712660715" watchObservedRunningTime="2025-10-07 22:27:37.511956885 +0000 UTC m=+5507.550538045" Oct 07 22:27:46 crc kubenswrapper[4926]: I1007 22:27:46.861610 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-wdx8h_7b429d99-24bf-4fee-81fb-eb1bc76d9788/nmstate-console-plugin/0.log" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.039513 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lswbm_8c97c92f-f0b8-48ec-b5f7-6594bd06c948/nmstate-handler/0.log" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.079612 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-mczl9_8daf114a-72f8-44e9-ac14-166bdf8e73cb/kube-rbac-proxy/0.log" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.130880 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-mczl9_8daf114a-72f8-44e9-ac14-166bdf8e73cb/nmstate-metrics/0.log" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.261592 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-vqnbn_c6ce449c-47a2-4bf7-9556-c3b72d502494/nmstate-operator/0.log" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.360032 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-r4nvj_aca14e9c-b601-4092-b7a5-72a80506b27e/nmstate-webhook/0.log" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.483757 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.554164 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4mpq"] Oct 07 22:27:47 crc kubenswrapper[4926]: I1007 22:27:47.804258 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t4mpq" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="registry-server" containerID="cri-o://29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73" gracePeriod=2 Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.332831 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.476883 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-catalog-content\") pod \"03a5135a-53b3-4594-adcb-44de3c6918f9\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.476976 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-utilities\") pod \"03a5135a-53b3-4594-adcb-44de3c6918f9\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.477018 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87m95\" (UniqueName: \"kubernetes.io/projected/03a5135a-53b3-4594-adcb-44de3c6918f9-kube-api-access-87m95\") pod \"03a5135a-53b3-4594-adcb-44de3c6918f9\" (UID: \"03a5135a-53b3-4594-adcb-44de3c6918f9\") " Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.477758 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-utilities" (OuterVolumeSpecName: "utilities") pod "03a5135a-53b3-4594-adcb-44de3c6918f9" (UID: "03a5135a-53b3-4594-adcb-44de3c6918f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.482887 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a5135a-53b3-4594-adcb-44de3c6918f9-kube-api-access-87m95" (OuterVolumeSpecName: "kube-api-access-87m95") pod "03a5135a-53b3-4594-adcb-44de3c6918f9" (UID: "03a5135a-53b3-4594-adcb-44de3c6918f9"). InnerVolumeSpecName "kube-api-access-87m95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.493965 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03a5135a-53b3-4594-adcb-44de3c6918f9" (UID: "03a5135a-53b3-4594-adcb-44de3c6918f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.579061 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.579093 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a5135a-53b3-4594-adcb-44de3c6918f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.579103 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87m95\" (UniqueName: \"kubernetes.io/projected/03a5135a-53b3-4594-adcb-44de3c6918f9-kube-api-access-87m95\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.816332 4926 generic.go:334] "Generic (PLEG): container finished" podID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerID="29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73" exitCode=0 Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.816384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerDied","Data":"29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73"} Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.816414 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4mpq" event={"ID":"03a5135a-53b3-4594-adcb-44de3c6918f9","Type":"ContainerDied","Data":"48d612f40a392e278186ff8a34456d2a0a44317f0d15123ccc7850b41bc68bb2"} Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.816435 4926 scope.go:117] "RemoveContainer" containerID="29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.816739 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4mpq" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.837925 4926 scope.go:117] "RemoveContainer" containerID="c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.843911 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4mpq"] Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.854752 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4mpq"] Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.862306 4926 scope.go:117] "RemoveContainer" containerID="89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.902128 4926 scope.go:117] "RemoveContainer" containerID="29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73" Oct 07 22:27:48 crc kubenswrapper[4926]: E1007 22:27:48.902511 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73\": container with ID starting with 29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73 not found: ID does not exist" containerID="29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.902541 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73"} err="failed to get container status \"29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73\": rpc error: code = NotFound desc = could not find container \"29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73\": container with ID starting with 29ad9ef7ff168ae1c41b3e8a69a2a7c1f1b77a76884f1370fb4481ceeb632d73 not found: ID does not exist" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.902562 4926 scope.go:117] "RemoveContainer" containerID="c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a" Oct 07 22:27:48 crc kubenswrapper[4926]: E1007 22:27:48.902846 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a\": container with ID starting with c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a not found: ID does not exist" containerID="c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.902973 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a"} err="failed to get container status \"c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a\": rpc error: code = NotFound desc = could not find container \"c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a\": container with ID starting with c2cfb02d42a88a2bd5d946f4bd8357e88e837376caeacdc4df0038f794417f4a not found: ID does not exist" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.903076 4926 scope.go:117] "RemoveContainer" containerID="89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14" Oct 07 22:27:48 crc kubenswrapper[4926]: E1007 22:27:48.903460 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14\": container with ID starting with 89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14 not found: ID does not exist" containerID="89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14" Oct 07 22:27:48 crc kubenswrapper[4926]: I1007 22:27:48.903491 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14"} err="failed to get container status \"89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14\": rpc error: code = NotFound desc = could not find container \"89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14\": container with ID starting with 89acc5eef76dcac6109e654ae0378107e6047fb5f14717d84abd7d5367d4ab14 not found: ID does not exist" Oct 07 22:27:50 crc kubenswrapper[4926]: I1007 22:27:50.697433 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" path="/var/lib/kubelet/pods/03a5135a-53b3-4594-adcb-44de3c6918f9/volumes" Oct 07 22:28:02 crc kubenswrapper[4926]: I1007 22:28:02.886564 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xdk65_17a8cd66-733c-4d61-83db-d0e56464b905/kube-rbac-proxy/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.107448 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xdk65_17a8cd66-733c-4d61-83db-d0e56464b905/controller/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.159505 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.318160 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.324466 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.359701 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.372389 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.570241 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.574288 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.600140 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.621499 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.796170 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.796738 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.801042 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/controller/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.801980 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:28:03 crc kubenswrapper[4926]: I1007 22:28:03.997958 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/frr-metrics/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.035277 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/kube-rbac-proxy/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.038766 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/kube-rbac-proxy-frr/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.288945 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-dh249_5c050d21-9d7e-4e41-9a8e-991580804642/frr-k8s-webhook-server/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.291778 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/reloader/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.577991 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74978ff576-w6wnd_fe0c800e-5525-4e10-8bb1-43a79192b2e8/manager/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.784335 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mg6kr_138368c1-eaa9-42ba-8b93-67bf6396ef26/kube-rbac-proxy/0.log" Oct 07 22:28:04 crc kubenswrapper[4926]: I1007 22:28:04.785374 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5f56c487b-tdcp8_472d3ee6-7ecc-471c-842c-94afa5bda368/webhook-server/0.log" Oct 07 22:28:05 crc kubenswrapper[4926]: I1007 22:28:05.500066 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mg6kr_138368c1-eaa9-42ba-8b93-67bf6396ef26/speaker/0.log" Oct 07 22:28:05 crc kubenswrapper[4926]: I1007 22:28:05.653352 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/frr/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.503903 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/util/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.607162 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/pull/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.612730 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/util/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.690953 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/pull/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.860905 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/pull/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.899541 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/util/0.log" Oct 07 22:28:20 crc kubenswrapper[4926]: I1007 22:28:20.933248 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/extract/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.077662 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/util/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.215164 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/util/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.225712 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/pull/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.296724 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/pull/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.439459 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/util/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.499344 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/extract/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.528515 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/pull/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.634232 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-utilities/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.782556 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-utilities/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.822683 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-content/0.log" Oct 07 22:28:21 crc kubenswrapper[4926]: I1007 22:28:21.839623 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-content/0.log" Oct 07 22:28:22 crc kubenswrapper[4926]: I1007 22:28:22.853069 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-utilities/0.log" Oct 07 22:28:22 crc kubenswrapper[4926]: I1007 22:28:22.889363 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-content/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.017675 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-utilities/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.278304 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-content/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.345204 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-content/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.356410 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-utilities/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.481047 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/registry-server/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.553805 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-utilities/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.577209 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-content/0.log" Oct 07 22:28:23 crc kubenswrapper[4926]: I1007 22:28:23.747734 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/util/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.324305 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/registry-server/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.419521 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/util/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.448519 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/pull/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.454299 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/pull/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.585430 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/util/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.602963 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/extract/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.623709 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/pull/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.670637 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rt2mf_1ffbcf7c-a254-485e-8da4-69c2892ca0e9/marketplace-operator/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.770996 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-utilities/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.926721 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-content/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.936457 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-utilities/0.log" Oct 07 22:28:24 crc kubenswrapper[4926]: I1007 22:28:24.961851 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-content/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.171864 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-utilities/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.191500 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-content/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.239979 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-utilities/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.301619 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/registry-server/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.388418 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-content/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.404668 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-utilities/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.412423 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-content/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.552980 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-content/0.log" Oct 07 22:28:25 crc kubenswrapper[4926]: I1007 22:28:25.585488 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-utilities/0.log" Oct 07 22:28:26 crc kubenswrapper[4926]: I1007 22:28:26.273597 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/registry-server/0.log" Oct 07 22:28:37 crc kubenswrapper[4926]: I1007 22:28:37.765379 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-hg4l9_0847160c-b3dd-4539-a141-d263faa41fd8/prometheus-operator/0.log" Oct 07 22:28:37 crc kubenswrapper[4926]: I1007 22:28:37.920615 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs_9f92dd43-5e91-41a3-bd84-8de059b4e9ee/prometheus-operator-admission-webhook/0.log" Oct 07 22:28:37 crc kubenswrapper[4926]: I1007 22:28:37.980647 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98_22bad45c-639a-4d86-aa08-1dd5f9b8ec5d/prometheus-operator-admission-webhook/0.log" Oct 07 22:28:38 crc kubenswrapper[4926]: I1007 22:28:38.100943 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-9pr8c_4d83bdfa-2009-43f5-93b1-cdc8e568020d/operator/0.log" Oct 07 22:28:38 crc kubenswrapper[4926]: I1007 22:28:38.166560 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-9vvdj_b4f527c2-fe8d-4da9-8477-2f02507a6178/perses-operator/0.log" Oct 07 22:28:44 crc kubenswrapper[4926]: E1007 22:28:44.865661 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.58:40428->38.102.83.58:42335: write tcp 38.102.83.58:40428->38.102.83.58:42335: write: broken pipe Oct 07 22:29:33 crc kubenswrapper[4926]: I1007 22:29:33.209331 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:29:33 crc kubenswrapper[4926]: I1007 22:29:33.209838 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.168988 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s"] Oct 07 22:30:00 crc kubenswrapper[4926]: E1007 22:30:00.177746 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="extract-content" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.177986 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="extract-content" Oct 07 22:30:00 crc kubenswrapper[4926]: E1007 22:30:00.178024 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="extract-utilities" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.178036 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="extract-utilities" Oct 07 22:30:00 crc kubenswrapper[4926]: E1007 22:30:00.178050 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="registry-server" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.178058 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="registry-server" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.178314 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a5135a-53b3-4594-adcb-44de3c6918f9" containerName="registry-server" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.179215 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.181695 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.184352 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.185285 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s"] Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.265746 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6de2dbe4-4826-4325-b53b-df5d2bd6474b-config-volume\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.265957 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6de2dbe4-4826-4325-b53b-df5d2bd6474b-secret-volume\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.266002 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-894ld\" (UniqueName: \"kubernetes.io/projected/6de2dbe4-4826-4325-b53b-df5d2bd6474b-kube-api-access-894ld\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.367768 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6de2dbe4-4826-4325-b53b-df5d2bd6474b-secret-volume\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.367850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-894ld\" (UniqueName: \"kubernetes.io/projected/6de2dbe4-4826-4325-b53b-df5d2bd6474b-kube-api-access-894ld\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.368795 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6de2dbe4-4826-4325-b53b-df5d2bd6474b-config-volume\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.370542 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6de2dbe4-4826-4325-b53b-df5d2bd6474b-config-volume\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.377895 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6de2dbe4-4826-4325-b53b-df5d2bd6474b-secret-volume\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.385571 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-894ld\" (UniqueName: \"kubernetes.io/projected/6de2dbe4-4826-4325-b53b-df5d2bd6474b-kube-api-access-894ld\") pod \"collect-profiles-29331270-c4w9s\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:00 crc kubenswrapper[4926]: I1007 22:30:00.518743 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:01 crc kubenswrapper[4926]: I1007 22:30:01.046638 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s"] Oct 07 22:30:01 crc kubenswrapper[4926]: I1007 22:30:01.265705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" event={"ID":"6de2dbe4-4826-4325-b53b-df5d2bd6474b","Type":"ContainerStarted","Data":"26eba387f354aa0387f9463b307917331e28d0b7fbeca8ce48ef74c6863582eb"} Oct 07 22:30:02 crc kubenswrapper[4926]: I1007 22:30:02.280168 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" event={"ID":"6de2dbe4-4826-4325-b53b-df5d2bd6474b","Type":"ContainerStarted","Data":"8c424415b462320b2feea6976c4269e9a7f592a338905bf332ce328f0e8e438c"} Oct 07 22:30:02 crc kubenswrapper[4926]: I1007 22:30:02.308451 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" podStartSLOduration=2.308429457 podStartE2EDuration="2.308429457s" podCreationTimestamp="2025-10-07 22:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:30:02.300948782 +0000 UTC m=+5652.339529942" watchObservedRunningTime="2025-10-07 22:30:02.308429457 +0000 UTC m=+5652.347010607" Oct 07 22:30:03 crc kubenswrapper[4926]: I1007 22:30:03.209931 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:30:03 crc kubenswrapper[4926]: I1007 22:30:03.210040 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:30:03 crc kubenswrapper[4926]: I1007 22:30:03.294657 4926 generic.go:334] "Generic (PLEG): container finished" podID="6de2dbe4-4826-4325-b53b-df5d2bd6474b" containerID="8c424415b462320b2feea6976c4269e9a7f592a338905bf332ce328f0e8e438c" exitCode=0 Oct 07 22:30:03 crc kubenswrapper[4926]: I1007 22:30:03.294728 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" event={"ID":"6de2dbe4-4826-4325-b53b-df5d2bd6474b","Type":"ContainerDied","Data":"8c424415b462320b2feea6976c4269e9a7f592a338905bf332ce328f0e8e438c"} Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.708398 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.766947 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6de2dbe4-4826-4325-b53b-df5d2bd6474b-config-volume\") pod \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.767021 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-894ld\" (UniqueName: \"kubernetes.io/projected/6de2dbe4-4826-4325-b53b-df5d2bd6474b-kube-api-access-894ld\") pod \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.767097 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6de2dbe4-4826-4325-b53b-df5d2bd6474b-secret-volume\") pod \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\" (UID: \"6de2dbe4-4826-4325-b53b-df5d2bd6474b\") " Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.767961 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6de2dbe4-4826-4325-b53b-df5d2bd6474b-config-volume" (OuterVolumeSpecName: "config-volume") pod "6de2dbe4-4826-4325-b53b-df5d2bd6474b" (UID: "6de2dbe4-4826-4325-b53b-df5d2bd6474b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.768332 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6de2dbe4-4826-4325-b53b-df5d2bd6474b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.774409 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6de2dbe4-4826-4325-b53b-df5d2bd6474b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6de2dbe4-4826-4325-b53b-df5d2bd6474b" (UID: "6de2dbe4-4826-4325-b53b-df5d2bd6474b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.774475 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6de2dbe4-4826-4325-b53b-df5d2bd6474b-kube-api-access-894ld" (OuterVolumeSpecName: "kube-api-access-894ld") pod "6de2dbe4-4826-4325-b53b-df5d2bd6474b" (UID: "6de2dbe4-4826-4325-b53b-df5d2bd6474b"). InnerVolumeSpecName "kube-api-access-894ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.870223 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-894ld\" (UniqueName: \"kubernetes.io/projected/6de2dbe4-4826-4325-b53b-df5d2bd6474b-kube-api-access-894ld\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:04 crc kubenswrapper[4926]: I1007 22:30:04.870271 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6de2dbe4-4826-4325-b53b-df5d2bd6474b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:05 crc kubenswrapper[4926]: I1007 22:30:05.316007 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" event={"ID":"6de2dbe4-4826-4325-b53b-df5d2bd6474b","Type":"ContainerDied","Data":"26eba387f354aa0387f9463b307917331e28d0b7fbeca8ce48ef74c6863582eb"} Oct 07 22:30:05 crc kubenswrapper[4926]: I1007 22:30:05.316059 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26eba387f354aa0387f9463b307917331e28d0b7fbeca8ce48ef74c6863582eb" Oct 07 22:30:05 crc kubenswrapper[4926]: I1007 22:30:05.316090 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-c4w9s" Oct 07 22:30:05 crc kubenswrapper[4926]: I1007 22:30:05.402800 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j"] Oct 07 22:30:05 crc kubenswrapper[4926]: I1007 22:30:05.418914 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331225-pkr6j"] Oct 07 22:30:06 crc kubenswrapper[4926]: I1007 22:30:06.699423 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833ec329-91dd-43e2-a699-7857055ead36" path="/var/lib/kubelet/pods/833ec329-91dd-43e2-a699-7857055ead36/volumes" Oct 07 22:30:06 crc kubenswrapper[4926]: I1007 22:30:06.937551 4926 scope.go:117] "RemoveContainer" containerID="539eae27cc239789eb0de7e5b7b6cea4b4234136945e47a0d656677380e22dee" Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.209493 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.215151 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.216326 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.219522 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.220178 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" gracePeriod=600 Oct 07 22:30:33 crc kubenswrapper[4926]: E1007 22:30:33.393420 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.639073 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" exitCode=0 Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.639127 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984"} Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.639166 4926 scope.go:117] "RemoveContainer" containerID="cb5490e43946c352966fc9267e77332a505cb51138c598a5313fc5b04210ef95" Oct 07 22:30:33 crc kubenswrapper[4926]: I1007 22:30:33.640072 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:30:33 crc kubenswrapper[4926]: E1007 22:30:33.640484 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:30:45 crc kubenswrapper[4926]: I1007 22:30:45.679939 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:30:45 crc kubenswrapper[4926]: E1007 22:30:45.680712 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:30:54 crc kubenswrapper[4926]: I1007 22:30:54.868514 4926 generic.go:334] "Generic (PLEG): container finished" podID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerID="9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232" exitCode=0 Oct 07 22:30:54 crc kubenswrapper[4926]: I1007 22:30:54.868703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-th6sx/must-gather-bn6kj" event={"ID":"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6","Type":"ContainerDied","Data":"9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232"} Oct 07 22:30:54 crc kubenswrapper[4926]: I1007 22:30:54.869839 4926 scope.go:117] "RemoveContainer" containerID="9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232" Oct 07 22:30:55 crc kubenswrapper[4926]: I1007 22:30:55.225521 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-th6sx_must-gather-bn6kj_7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6/gather/0.log" Oct 07 22:30:57 crc kubenswrapper[4926]: I1007 22:30:57.681512 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:30:57 crc kubenswrapper[4926]: E1007 22:30:57.682088 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:31:05 crc kubenswrapper[4926]: I1007 22:31:05.359805 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-th6sx/must-gather-bn6kj"] Oct 07 22:31:05 crc kubenswrapper[4926]: I1007 22:31:05.363074 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-th6sx/must-gather-bn6kj" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="copy" containerID="cri-o://ce307b5cb97e5f4916afcb8025619cc9cdeccd1e6d79fe0eb61a9abda4b5b42b" gracePeriod=2 Oct 07 22:31:05 crc kubenswrapper[4926]: I1007 22:31:05.389731 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-th6sx/must-gather-bn6kj"] Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.071291 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-th6sx_must-gather-bn6kj_7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6/copy/0.log" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.073889 4926 generic.go:334] "Generic (PLEG): container finished" podID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerID="ce307b5cb97e5f4916afcb8025619cc9cdeccd1e6d79fe0eb61a9abda4b5b42b" exitCode=143 Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.295573 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-th6sx_must-gather-bn6kj_7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6/copy/0.log" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.296037 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.380149 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-must-gather-output\") pod \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.380334 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5tks\" (UniqueName: \"kubernetes.io/projected/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-kube-api-access-d5tks\") pod \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\" (UID: \"7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6\") " Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.396419 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-kube-api-access-d5tks" (OuterVolumeSpecName: "kube-api-access-d5tks") pod "7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" (UID: "7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6"). InnerVolumeSpecName "kube-api-access-d5tks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.483592 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5tks\" (UniqueName: \"kubernetes.io/projected/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-kube-api-access-d5tks\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.610883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" (UID: "7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.687973 4926 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:06 crc kubenswrapper[4926]: I1007 22:31:06.691692 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" path="/var/lib/kubelet/pods/7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6/volumes" Oct 07 22:31:07 crc kubenswrapper[4926]: I1007 22:31:07.086762 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-th6sx_must-gather-bn6kj_7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6/copy/0.log" Oct 07 22:31:07 crc kubenswrapper[4926]: I1007 22:31:07.087462 4926 scope.go:117] "RemoveContainer" containerID="ce307b5cb97e5f4916afcb8025619cc9cdeccd1e6d79fe0eb61a9abda4b5b42b" Oct 07 22:31:07 crc kubenswrapper[4926]: I1007 22:31:07.087573 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-th6sx/must-gather-bn6kj" Oct 07 22:31:07 crc kubenswrapper[4926]: I1007 22:31:07.120432 4926 scope.go:117] "RemoveContainer" containerID="9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232" Oct 07 22:31:07 crc kubenswrapper[4926]: I1007 22:31:07.138113 4926 scope.go:117] "RemoveContainer" containerID="9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232" Oct 07 22:31:07 crc kubenswrapper[4926]: E1007 22:31:07.235903 4926 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_gather_must-gather-bn6kj_openshift-must-gather-th6sx_7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6_0 in pod sandbox 898620ece972cda47cbfa0f4af561cc452074a0801f06a57d73aa5477222680a: identifier is not a container" containerID="9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232" Oct 07 22:31:07 crc kubenswrapper[4926]: E1007 22:31:07.235984 4926 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_gather_must-gather-bn6kj_openshift-must-gather-th6sx_7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6_0 in pod sandbox 898620ece972cda47cbfa0f4af561cc452074a0801f06a57d73aa5477222680a: identifier is not a container" containerID="9ebc33b2dd53654e75e50ef9f83aec863f7d0efeed6f1710f3fa384d7554e232" Oct 07 22:31:09 crc kubenswrapper[4926]: I1007 22:31:09.678546 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:31:09 crc kubenswrapper[4926]: E1007 22:31:09.679483 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.203717 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lqg7x"] Oct 07 22:31:19 crc kubenswrapper[4926]: E1007 22:31:19.206322 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="gather" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.206341 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="gather" Oct 07 22:31:19 crc kubenswrapper[4926]: E1007 22:31:19.206352 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de2dbe4-4826-4325-b53b-df5d2bd6474b" containerName="collect-profiles" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.206358 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de2dbe4-4826-4325-b53b-df5d2bd6474b" containerName="collect-profiles" Oct 07 22:31:19 crc kubenswrapper[4926]: E1007 22:31:19.206374 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="copy" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.206380 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="copy" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.206564 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="gather" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.206590 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9a4b58-6d90-4dc4-a1ab-17d046c3adf6" containerName="copy" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.206611 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de2dbe4-4826-4325-b53b-df5d2bd6474b" containerName="collect-profiles" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.207942 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.225683 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lqg7x"] Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.293800 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-utilities\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.294087 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsfjp\" (UniqueName: \"kubernetes.io/projected/cea701a8-efaf-4991-8827-0d1990fda6cc-kube-api-access-lsfjp\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.294229 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-catalog-content\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.396277 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-utilities\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.396349 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsfjp\" (UniqueName: \"kubernetes.io/projected/cea701a8-efaf-4991-8827-0d1990fda6cc-kube-api-access-lsfjp\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.396381 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-catalog-content\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.396901 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-utilities\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.396912 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-catalog-content\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.426047 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsfjp\" (UniqueName: \"kubernetes.io/projected/cea701a8-efaf-4991-8827-0d1990fda6cc-kube-api-access-lsfjp\") pod \"redhat-operators-lqg7x\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:19 crc kubenswrapper[4926]: I1007 22:31:19.527025 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:20 crc kubenswrapper[4926]: I1007 22:31:20.071469 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lqg7x"] Oct 07 22:31:20 crc kubenswrapper[4926]: I1007 22:31:20.235718 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqg7x" event={"ID":"cea701a8-efaf-4991-8827-0d1990fda6cc","Type":"ContainerStarted","Data":"1f530d60c300b8f255077418b2f719730077bc10d5831c10dad0c46248b25e7e"} Oct 07 22:31:21 crc kubenswrapper[4926]: I1007 22:31:21.249901 4926 generic.go:334] "Generic (PLEG): container finished" podID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerID="1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7" exitCode=0 Oct 07 22:31:21 crc kubenswrapper[4926]: I1007 22:31:21.249982 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqg7x" event={"ID":"cea701a8-efaf-4991-8827-0d1990fda6cc","Type":"ContainerDied","Data":"1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7"} Oct 07 22:31:21 crc kubenswrapper[4926]: I1007 22:31:21.253555 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:31:22 crc kubenswrapper[4926]: I1007 22:31:22.680165 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:31:22 crc kubenswrapper[4926]: E1007 22:31:22.680903 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:31:23 crc kubenswrapper[4926]: I1007 22:31:23.273845 4926 generic.go:334] "Generic (PLEG): container finished" podID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerID="8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe" exitCode=0 Oct 07 22:31:23 crc kubenswrapper[4926]: I1007 22:31:23.273898 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqg7x" event={"ID":"cea701a8-efaf-4991-8827-0d1990fda6cc","Type":"ContainerDied","Data":"8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe"} Oct 07 22:31:24 crc kubenswrapper[4926]: I1007 22:31:24.283537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqg7x" event={"ID":"cea701a8-efaf-4991-8827-0d1990fda6cc","Type":"ContainerStarted","Data":"88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9"} Oct 07 22:31:24 crc kubenswrapper[4926]: I1007 22:31:24.301872 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lqg7x" podStartSLOduration=2.652690697 podStartE2EDuration="5.301856008s" podCreationTimestamp="2025-10-07 22:31:19 +0000 UTC" firstStartedPulling="2025-10-07 22:31:21.253180716 +0000 UTC m=+5731.291761876" lastFinishedPulling="2025-10-07 22:31:23.902346037 +0000 UTC m=+5733.940927187" observedRunningTime="2025-10-07 22:31:24.297596725 +0000 UTC m=+5734.336177875" watchObservedRunningTime="2025-10-07 22:31:24.301856008 +0000 UTC m=+5734.340437178" Oct 07 22:31:29 crc kubenswrapper[4926]: I1007 22:31:29.527187 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:29 crc kubenswrapper[4926]: I1007 22:31:29.527670 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:29 crc kubenswrapper[4926]: I1007 22:31:29.581050 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:30 crc kubenswrapper[4926]: I1007 22:31:30.418667 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:30 crc kubenswrapper[4926]: I1007 22:31:30.480294 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lqg7x"] Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.336531 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6c9g9/must-gather-sschm"] Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.338948 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.342438 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6c9g9"/"openshift-service-ca.crt" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.342600 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6c9g9"/"kube-root-ca.crt" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.349176 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6c9g9/must-gather-sschm"] Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.384630 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lqg7x" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="registry-server" containerID="cri-o://88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9" gracePeriod=2 Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.481228 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5clmg\" (UniqueName: \"kubernetes.io/projected/2243e5f5-ea2c-4456-be75-6f10706aaa1f-kube-api-access-5clmg\") pod \"must-gather-sschm\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.481339 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output\") pod \"must-gather-sschm\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.623642 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5clmg\" (UniqueName: \"kubernetes.io/projected/2243e5f5-ea2c-4456-be75-6f10706aaa1f-kube-api-access-5clmg\") pod \"must-gather-sschm\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.624093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output\") pod \"must-gather-sschm\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.624836 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output\") pod \"must-gather-sschm\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.648864 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5clmg\" (UniqueName: \"kubernetes.io/projected/2243e5f5-ea2c-4456-be75-6f10706aaa1f-kube-api-access-5clmg\") pod \"must-gather-sschm\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:32 crc kubenswrapper[4926]: I1007 22:31:32.670594 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.035093 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.201878 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-catalog-content\") pod \"cea701a8-efaf-4991-8827-0d1990fda6cc\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.201993 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-utilities\") pod \"cea701a8-efaf-4991-8827-0d1990fda6cc\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.202152 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsfjp\" (UniqueName: \"kubernetes.io/projected/cea701a8-efaf-4991-8827-0d1990fda6cc-kube-api-access-lsfjp\") pod \"cea701a8-efaf-4991-8827-0d1990fda6cc\" (UID: \"cea701a8-efaf-4991-8827-0d1990fda6cc\") " Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.203260 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-utilities" (OuterVolumeSpecName: "utilities") pod "cea701a8-efaf-4991-8827-0d1990fda6cc" (UID: "cea701a8-efaf-4991-8827-0d1990fda6cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.206844 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea701a8-efaf-4991-8827-0d1990fda6cc-kube-api-access-lsfjp" (OuterVolumeSpecName: "kube-api-access-lsfjp") pod "cea701a8-efaf-4991-8827-0d1990fda6cc" (UID: "cea701a8-efaf-4991-8827-0d1990fda6cc"). InnerVolumeSpecName "kube-api-access-lsfjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.305470 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsfjp\" (UniqueName: \"kubernetes.io/projected/cea701a8-efaf-4991-8827-0d1990fda6cc-kube-api-access-lsfjp\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.305532 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:33 crc kubenswrapper[4926]: W1007 22:31:33.337338 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2243e5f5_ea2c_4456_be75_6f10706aaa1f.slice/crio-b055124dc058e942fc4c189f8cd82ea4d2a74d48dd05d65840df09ae60c7281c WatchSource:0}: Error finding container b055124dc058e942fc4c189f8cd82ea4d2a74d48dd05d65840df09ae60c7281c: Status 404 returned error can't find the container with id b055124dc058e942fc4c189f8cd82ea4d2a74d48dd05d65840df09ae60c7281c Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.337415 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6c9g9/must-gather-sschm"] Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.396624 4926 generic.go:334] "Generic (PLEG): container finished" podID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerID="88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9" exitCode=0 Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.396791 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqg7x" event={"ID":"cea701a8-efaf-4991-8827-0d1990fda6cc","Type":"ContainerDied","Data":"88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9"} Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.396912 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqg7x" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.396972 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqg7x" event={"ID":"cea701a8-efaf-4991-8827-0d1990fda6cc","Type":"ContainerDied","Data":"1f530d60c300b8f255077418b2f719730077bc10d5831c10dad0c46248b25e7e"} Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.397108 4926 scope.go:117] "RemoveContainer" containerID="88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.399089 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/must-gather-sschm" event={"ID":"2243e5f5-ea2c-4456-be75-6f10706aaa1f","Type":"ContainerStarted","Data":"b055124dc058e942fc4c189f8cd82ea4d2a74d48dd05d65840df09ae60c7281c"} Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.430412 4926 scope.go:117] "RemoveContainer" containerID="8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.457848 4926 scope.go:117] "RemoveContainer" containerID="1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.491340 4926 scope.go:117] "RemoveContainer" containerID="88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9" Oct 07 22:31:33 crc kubenswrapper[4926]: E1007 22:31:33.491817 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9\": container with ID starting with 88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9 not found: ID does not exist" containerID="88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.491851 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9"} err="failed to get container status \"88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9\": rpc error: code = NotFound desc = could not find container \"88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9\": container with ID starting with 88bbe8c90a4730fa9418723486b2516910411a07553d9fa140cbceb8a6ea19a9 not found: ID does not exist" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.491877 4926 scope.go:117] "RemoveContainer" containerID="8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe" Oct 07 22:31:33 crc kubenswrapper[4926]: E1007 22:31:33.492274 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe\": container with ID starting with 8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe not found: ID does not exist" containerID="8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.492347 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe"} err="failed to get container status \"8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe\": rpc error: code = NotFound desc = could not find container \"8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe\": container with ID starting with 8721a3b623dba28c9d6e28122045d126d253728de2982098936ccd0ec3bd42fe not found: ID does not exist" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.492390 4926 scope.go:117] "RemoveContainer" containerID="1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7" Oct 07 22:31:33 crc kubenswrapper[4926]: E1007 22:31:33.493026 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7\": container with ID starting with 1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7 not found: ID does not exist" containerID="1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.493091 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7"} err="failed to get container status \"1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7\": rpc error: code = NotFound desc = could not find container \"1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7\": container with ID starting with 1a546f8d34048309bd039a52cc0faa21a65f1ca1f1da2793b16c25374ee5fff7 not found: ID does not exist" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.854422 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cea701a8-efaf-4991-8827-0d1990fda6cc" (UID: "cea701a8-efaf-4991-8827-0d1990fda6cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:33 crc kubenswrapper[4926]: I1007 22:31:33.921044 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea701a8-efaf-4991-8827-0d1990fda6cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:34 crc kubenswrapper[4926]: I1007 22:31:34.038366 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lqg7x"] Oct 07 22:31:34 crc kubenswrapper[4926]: I1007 22:31:34.048617 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lqg7x"] Oct 07 22:31:34 crc kubenswrapper[4926]: I1007 22:31:34.413600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/must-gather-sschm" event={"ID":"2243e5f5-ea2c-4456-be75-6f10706aaa1f","Type":"ContainerStarted","Data":"562eeb5031d037ce8862de91b9c9e07b8d941d0bac3fe902032947593becdd95"} Oct 07 22:31:34 crc kubenswrapper[4926]: I1007 22:31:34.414083 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/must-gather-sschm" event={"ID":"2243e5f5-ea2c-4456-be75-6f10706aaa1f","Type":"ContainerStarted","Data":"2a59cffc4f7644ce4fd07a44aecfefa5bde6b30836fa72b42f94ee5bc4172e15"} Oct 07 22:31:34 crc kubenswrapper[4926]: I1007 22:31:34.435321 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6c9g9/must-gather-sschm" podStartSLOduration=2.435305533 podStartE2EDuration="2.435305533s" podCreationTimestamp="2025-10-07 22:31:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:34.432218094 +0000 UTC m=+5744.470799254" watchObservedRunningTime="2025-10-07 22:31:34.435305533 +0000 UTC m=+5744.473886683" Oct 07 22:31:34 crc kubenswrapper[4926]: I1007 22:31:34.690419 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" path="/var/lib/kubelet/pods/cea701a8-efaf-4991-8827-0d1990fda6cc/volumes" Oct 07 22:31:35 crc kubenswrapper[4926]: I1007 22:31:35.679384 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:31:35 crc kubenswrapper[4926]: E1007 22:31:35.680602 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.545387 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-dt5qs"] Oct 07 22:31:37 crc kubenswrapper[4926]: E1007 22:31:37.546681 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="extract-utilities" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.546703 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="extract-utilities" Oct 07 22:31:37 crc kubenswrapper[4926]: E1007 22:31:37.546720 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="registry-server" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.546727 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="registry-server" Oct 07 22:31:37 crc kubenswrapper[4926]: E1007 22:31:37.546773 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="extract-content" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.546782 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="extract-content" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.547019 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea701a8-efaf-4991-8827-0d1990fda6cc" containerName="registry-server" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.547861 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.550539 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6c9g9"/"default-dockercfg-tccpd" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.693353 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-host\") pod \"crc-debug-dt5qs\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.693609 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9hsj\" (UniqueName: \"kubernetes.io/projected/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-kube-api-access-b9hsj\") pod \"crc-debug-dt5qs\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.796083 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9hsj\" (UniqueName: \"kubernetes.io/projected/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-kube-api-access-b9hsj\") pod \"crc-debug-dt5qs\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.796307 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-host\") pod \"crc-debug-dt5qs\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.796524 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-host\") pod \"crc-debug-dt5qs\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.816899 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9hsj\" (UniqueName: \"kubernetes.io/projected/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-kube-api-access-b9hsj\") pod \"crc-debug-dt5qs\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: I1007 22:31:37.868870 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:31:37 crc kubenswrapper[4926]: W1007 22:31:37.912665 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9764a0e8_f3f4_4206_9dc9_14a55ebf4344.slice/crio-54ce50dfe273ca439a3ab371b87398b5d701fc4459b954613c8ec8e2b1f0f981 WatchSource:0}: Error finding container 54ce50dfe273ca439a3ab371b87398b5d701fc4459b954613c8ec8e2b1f0f981: Status 404 returned error can't find the container with id 54ce50dfe273ca439a3ab371b87398b5d701fc4459b954613c8ec8e2b1f0f981 Oct 07 22:31:38 crc kubenswrapper[4926]: I1007 22:31:38.459675 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" event={"ID":"9764a0e8-f3f4-4206-9dc9-14a55ebf4344","Type":"ContainerStarted","Data":"90f32ab8e60f2834c1abfb62143c6dfb3d27313bbba119ae4e2e5d77dd1f7360"} Oct 07 22:31:38 crc kubenswrapper[4926]: I1007 22:31:38.460359 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" event={"ID":"9764a0e8-f3f4-4206-9dc9-14a55ebf4344","Type":"ContainerStarted","Data":"54ce50dfe273ca439a3ab371b87398b5d701fc4459b954613c8ec8e2b1f0f981"} Oct 07 22:31:38 crc kubenswrapper[4926]: I1007 22:31:38.487564 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" podStartSLOduration=1.4875314099999999 podStartE2EDuration="1.48753141s" podCreationTimestamp="2025-10-07 22:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:38.482813234 +0000 UTC m=+5748.521394394" watchObservedRunningTime="2025-10-07 22:31:38.48753141 +0000 UTC m=+5748.526112560" Oct 07 22:31:50 crc kubenswrapper[4926]: I1007 22:31:50.686258 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:31:50 crc kubenswrapper[4926]: E1007 22:31:50.687040 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:32:02 crc kubenswrapper[4926]: I1007 22:32:02.679210 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:32:02 crc kubenswrapper[4926]: E1007 22:32:02.680009 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:32:16 crc kubenswrapper[4926]: I1007 22:32:16.678827 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:32:16 crc kubenswrapper[4926]: E1007 22:32:16.679525 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:32:29 crc kubenswrapper[4926]: I1007 22:32:29.679469 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:32:29 crc kubenswrapper[4926]: E1007 22:32:29.680227 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:32:42 crc kubenswrapper[4926]: I1007 22:32:42.688645 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:32:42 crc kubenswrapper[4926]: E1007 22:32:42.689618 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.204278 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6696d67754-xn5b4_d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c/barbican-api/0.log" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.325061 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6696d67754-xn5b4_d1e8f987-f0b3-4fdd-bbf3-2a1ce9fe0a6c/barbican-api-log/0.log" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.445167 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6774c5f56-pjjjg_5badd158-2c74-425b-bc80-95ebba967905/barbican-keystone-listener/0.log" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.584650 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6774c5f56-pjjjg_5badd158-2c74-425b-bc80-95ebba967905/barbican-keystone-listener-log/0.log" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.642227 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-86dd45b86f-x4sv6_afa6dc20-30d9-48f1-886f-6b436a2f0678/barbican-worker/0.log" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.806163 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-86dd45b86f-x4sv6_afa6dc20-30d9-48f1-886f-6b436a2f0678/barbican-worker-log/0.log" Oct 07 22:32:55 crc kubenswrapper[4926]: I1007 22:32:55.914735 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-qw6pn_9612131a-d029-4ed0-a6a0-f2562d6b34d2/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.142014 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/ceilometer-notification-agent/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.208963 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/ceilometer-central-agent/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.233998 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/proxy-httpd/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.326258 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_69db6fef-77ff-4e55-9a3c-2df537ae1632/sg-core/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.565085 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_aabe891b-eda8-4359-b8c8-d4d99e309148/cinder-api-log/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.678410 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:32:56 crc kubenswrapper[4926]: E1007 22:32:56.678636 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.774044 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_aabe891b-eda8-4359-b8c8-d4d99e309148/cinder-api/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.842210 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_348d1b67-e2d4-4699-9890-3e3befa0133c/cinder-scheduler/0.log" Oct 07 22:32:56 crc kubenswrapper[4926]: I1007 22:32:56.935777 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_348d1b67-e2d4-4699-9890-3e3befa0133c/probe/0.log" Oct 07 22:32:57 crc kubenswrapper[4926]: I1007 22:32:57.110495 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-p2jbm_67f58dbd-5fcb-47a6-a904-7e3c5232dd1e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:57 crc kubenswrapper[4926]: I1007 22:32:57.250938 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lhm27_3f3a65d1-ea19-4587-a483-ba006c9d4660/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:57 crc kubenswrapper[4926]: I1007 22:32:57.405687 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-rllwb_2e0b270f-6b8e-47a4-b0ea-070dc5d2bd47/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:57 crc kubenswrapper[4926]: I1007 22:32:57.598861 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c8d6588b7-27rms_f8f8c2fb-1675-4abc-8ae2-dbe68370e347/init/0.log" Oct 07 22:32:57 crc kubenswrapper[4926]: I1007 22:32:57.764909 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c8d6588b7-27rms_f8f8c2fb-1675-4abc-8ae2-dbe68370e347/init/0.log" Oct 07 22:32:57 crc kubenswrapper[4926]: I1007 22:32:57.960209 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c8d6588b7-27rms_f8f8c2fb-1675-4abc-8ae2-dbe68370e347/dnsmasq-dns/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.028609 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pv8dq_ab42f719-a43c-425e-b0b3-aff4f812aefb/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.182051 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_15c1a57a-dd1c-4955-ab28-86f0d09faffb/glance-log/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.197969 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_15c1a57a-dd1c-4955-ab28-86f0d09faffb/glance-httpd/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.473277 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e4b6a533-2f48-428f-b735-3ceac2e2d7c7/glance-httpd/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.519537 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e4b6a533-2f48-428f-b735-3ceac2e2d7c7/glance-log/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.818891 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-858b4c475d-fcp7q_c42be0d4-2ced-4d32-84f4-49389e2a0e82/horizon/0.log" Oct 07 22:32:58 crc kubenswrapper[4926]: I1007 22:32:58.911232 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4wtxf_52cd3a5e-b86a-49c9-84fd-387676280d79/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:59 crc kubenswrapper[4926]: I1007 22:32:59.033965 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-m9fn4_d64881a0-3a23-4075-94c1-46f52c09e5ef/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:32:59 crc kubenswrapper[4926]: I1007 22:32:59.285299 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-858b4c475d-fcp7q_c42be0d4-2ced-4d32-84f4-49389e2a0e82/horizon-log/0.log" Oct 07 22:32:59 crc kubenswrapper[4926]: I1007 22:32:59.374501 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29331241-qg8pm_ab4d73e0-30e0-4388-b7dc-28c97399930d/keystone-cron/0.log" Oct 07 22:32:59 crc kubenswrapper[4926]: I1007 22:32:59.617704 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b8c45bfd-abbf-432e-bb7f-70739b19c587/kube-state-metrics/0.log" Oct 07 22:32:59 crc kubenswrapper[4926]: I1007 22:32:59.672813 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-647f4cc4bb-zkt6b_434a0a19-516e-4685-be72-36976397fa42/keystone-api/0.log" Oct 07 22:32:59 crc kubenswrapper[4926]: I1007 22:32:59.768022 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-j7hqr_4c8adb00-aa62-4d33-9b2f-a9b19b57e893/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:00 crc kubenswrapper[4926]: I1007 22:33:00.216888 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-b64c68dfc-k895v_c8be70c1-8342-414d-ba4b-098d40277bf6/neutron-api/0.log" Oct 07 22:33:00 crc kubenswrapper[4926]: I1007 22:33:00.273289 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-b64c68dfc-k895v_c8be70c1-8342-414d-ba4b-098d40277bf6/neutron-httpd/0.log" Oct 07 22:33:00 crc kubenswrapper[4926]: I1007 22:33:00.424033 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nbmvk_bd775971-aba6-499a-b63f-ee121ded14d4/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:01 crc kubenswrapper[4926]: I1007 22:33:01.329389 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b4601f7d-946c-4632-9778-44f70cafc757/nova-cell0-conductor-conductor/0.log" Oct 07 22:33:02 crc kubenswrapper[4926]: I1007 22:33:02.156250 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2f6780f7-788b-42cf-937e-750a222de004/nova-cell1-conductor-conductor/0.log" Oct 07 22:33:02 crc kubenswrapper[4926]: I1007 22:33:02.198658 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_08044ebc-ce53-4476-9884-83792098080d/nova-api-log/0.log" Oct 07 22:33:02 crc kubenswrapper[4926]: I1007 22:33:02.481602 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_08044ebc-ce53-4476-9884-83792098080d/nova-api-api/0.log" Oct 07 22:33:02 crc kubenswrapper[4926]: I1007 22:33:02.557176 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6e980975-58c0-4255-bc3e-df92773323cf/nova-cell1-novncproxy-novncproxy/0.log" Oct 07 22:33:02 crc kubenswrapper[4926]: I1007 22:33:02.717944 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-t7hkb_0273b882-c15f-4ca8-b786-60fb42c205ab/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:02 crc kubenswrapper[4926]: I1007 22:33:02.865903 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4496af78-6ebe-4aec-a04e-dcf4e8f446eb/nova-metadata-log/0.log" Oct 07 22:33:03 crc kubenswrapper[4926]: I1007 22:33:03.423019 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c5fe2fcd-3215-4278-90fc-3adc5cec14fe/nova-scheduler-scheduler/0.log" Oct 07 22:33:03 crc kubenswrapper[4926]: I1007 22:33:03.494559 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d169a79-a57d-458f-a790-3623f6dd0885/mysql-bootstrap/0.log" Oct 07 22:33:03 crc kubenswrapper[4926]: I1007 22:33:03.632543 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d169a79-a57d-458f-a790-3623f6dd0885/mysql-bootstrap/0.log" Oct 07 22:33:03 crc kubenswrapper[4926]: I1007 22:33:03.719498 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d169a79-a57d-458f-a790-3623f6dd0885/galera/0.log" Oct 07 22:33:03 crc kubenswrapper[4926]: I1007 22:33:03.952460 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7bf0bb92-3a4f-44ae-be03-e260cb246adb/mysql-bootstrap/0.log" Oct 07 22:33:04 crc kubenswrapper[4926]: I1007 22:33:04.094069 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7bf0bb92-3a4f-44ae-be03-e260cb246adb/mysql-bootstrap/0.log" Oct 07 22:33:04 crc kubenswrapper[4926]: I1007 22:33:04.178858 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7bf0bb92-3a4f-44ae-be03-e260cb246adb/galera/0.log" Oct 07 22:33:04 crc kubenswrapper[4926]: I1007 22:33:04.355084 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2f161d5b-f135-4e3a-81b3-0ab0ea2f7e15/openstackclient/0.log" Oct 07 22:33:04 crc kubenswrapper[4926]: I1007 22:33:04.565978 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6qkqt_42ae43bc-ecb7-4cb8-95e5-394f7bc04dc2/ovn-controller/0.log" Oct 07 22:33:04 crc kubenswrapper[4926]: I1007 22:33:04.760858 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5g8lv_5a9488ef-d520-4ff8-9eb2-8de1c72f03ce/openstack-network-exporter/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.222639 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4496af78-6ebe-4aec-a04e-dcf4e8f446eb/nova-metadata-metadata/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.247963 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovsdb-server-init/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.395848 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovsdb-server-init/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.490548 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovsdb-server/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.723413 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zg8nq_188a9695-679b-46c5-ac4b-26f71c9b4110/ovs-vswitchd/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.758937 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cwx7z_bcecc1ad-1ce0-4b1a-8861-fc7d37276b12/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.927790 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee1a017c-a9f5-497d-9b4d-80f6f776be50/openstack-network-exporter/0.log" Oct 07 22:33:05 crc kubenswrapper[4926]: I1007 22:33:05.977670 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee1a017c-a9f5-497d-9b4d-80f6f776be50/ovn-northd/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.120967 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b2704dcf-cd03-4f91-af4e-ea708f3c0534/openstack-network-exporter/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.192989 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b2704dcf-cd03-4f91-af4e-ea708f3c0534/ovsdbserver-nb/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.388846 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fdc18634-337d-4562-a6e5-a20df1f53028/ovsdbserver-sb/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.392086 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fdc18634-337d-4562-a6e5-a20df1f53028/openstack-network-exporter/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.756436 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-849bb7d7cd-xqmg2_1af8a1d9-b36b-4f59-905c-12d2245ae3cf/placement-api/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.866960 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-849bb7d7cd-xqmg2_1af8a1d9-b36b-4f59-905c-12d2245ae3cf/placement-log/0.log" Oct 07 22:33:06 crc kubenswrapper[4926]: I1007 22:33:06.972018 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/init-config-reloader/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.151148 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/config-reloader/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.171514 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/init-config-reloader/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.224280 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/prometheus/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.330642 4926 scope.go:117] "RemoveContainer" containerID="605ab15b8f73526a347a3fc976ae82287f612974090f32b79f9b144552d0f1b3" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.446421 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_47ab84ef-8080-42f6-9467-6d1439275f65/thanos-sidecar/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.516500 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d5d417c9-ac1d-4b58-9506-bc6dd17a71b1/setup-container/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.721457 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d5d417c9-ac1d-4b58-9506-bc6dd17a71b1/setup-container/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.757064 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d5d417c9-ac1d-4b58-9506-bc6dd17a71b1/rabbitmq/0.log" Oct 07 22:33:07 crc kubenswrapper[4926]: I1007 22:33:07.926146 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_5b0d7388-05f4-4df6-b093-c9157a4658b0/setup-container/0.log" Oct 07 22:33:08 crc kubenswrapper[4926]: I1007 22:33:08.240074 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_5b0d7388-05f4-4df6-b093-c9157a4658b0/rabbitmq/0.log" Oct 07 22:33:08 crc kubenswrapper[4926]: I1007 22:33:08.258091 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_5b0d7388-05f4-4df6-b093-c9157a4658b0/setup-container/0.log" Oct 07 22:33:08 crc kubenswrapper[4926]: I1007 22:33:08.572117 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9052310e-1ddc-4b9b-a3bc-ced3dc97ab21/setup-container/0.log" Oct 07 22:33:08 crc kubenswrapper[4926]: I1007 22:33:08.792175 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9052310e-1ddc-4b9b-a3bc-ced3dc97ab21/rabbitmq/0.log" Oct 07 22:33:08 crc kubenswrapper[4926]: I1007 22:33:08.795239 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9052310e-1ddc-4b9b-a3bc-ced3dc97ab21/setup-container/0.log" Oct 07 22:33:09 crc kubenswrapper[4926]: I1007 22:33:09.082013 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-hkpfd_28a80874-caa4-48eb-9cbc-999e2db615fb/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:09 crc kubenswrapper[4926]: I1007 22:33:09.146133 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-hnftc_b581b351-be97-4417-9c19-58e962bf821e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:09 crc kubenswrapper[4926]: I1007 22:33:09.346778 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7z4qc_b9c1704c-58c8-407c-bafc-c23a19cf900b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:09 crc kubenswrapper[4926]: I1007 22:33:09.575983 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kbwzm_67068a37-7d4a-47ca-9971-89e1e7a49a25/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:09 crc kubenswrapper[4926]: I1007 22:33:09.674837 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-5rwsd_69fe71d5-112a-45aa-a043-8629073ec0c2/ssh-known-hosts-edpm-deployment/0.log" Oct 07 22:33:09 crc kubenswrapper[4926]: I1007 22:33:09.901716 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55b995649c-qwlp9_28011eca-ae6f-42da-9e6b-27289eb19b2e/proxy-server/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.023495 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55b995649c-qwlp9_28011eca-ae6f-42da-9e6b-27289eb19b2e/proxy-httpd/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.129413 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-fz74j_f314e4d4-9c99-4880-ab6e-a2eb04e2ae5f/swift-ring-rebalance/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.302403 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-auditor/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.359809 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-reaper/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.519628 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-replicator/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.520527 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/account-server/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.619876 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-auditor/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.725651 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-server/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.804886 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-replicator/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.855448 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/container-updater/0.log" Oct 07 22:33:10 crc kubenswrapper[4926]: I1007 22:33:10.952266 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-auditor/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.019720 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-expirer/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.120654 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-replicator/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.174031 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_038e2e94-c12a-4d54-8b48-96d08012ab97/memcached/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.194949 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-server/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.222346 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/object-updater/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.286292 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/rsync/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.338576 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8671e7ac-5c56-4b89-9d72-386b75d12ec1/swift-recon-cron/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.451145 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v99xb_013ef09a-4927-4ca6-9beb-0537614a9187/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.545493 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_25a6326c-d6df-46b2-b646-0c37aa3bb0ed/tempest-tests-tempest-tests-runner/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.616923 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_56b825d8-95d6-44ae-bad6-3f498d42f8db/test-operator-logs-container/0.log" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.679017 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:33:11 crc kubenswrapper[4926]: E1007 22:33:11.679291 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:33:11 crc kubenswrapper[4926]: I1007 22:33:11.778974 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4z9tx_a94c25aa-d7cb-4329-8dbc-d52c8b511cf3/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 22:33:12 crc kubenswrapper[4926]: I1007 22:33:12.646131 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_7d2feaba-0aff-4563-94f6-939db1594277/watcher-applier/0.log" Oct 07 22:33:12 crc kubenswrapper[4926]: I1007 22:33:12.993784 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_3d69b703-1e74-4e92-943d-6c73f029020e/watcher-api-log/0.log" Oct 07 22:33:15 crc kubenswrapper[4926]: I1007 22:33:15.084309 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_d9756c29-8c2e-49c9-8bf2-f9a51710d7d7/watcher-decision-engine/0.log" Oct 07 22:33:15 crc kubenswrapper[4926]: I1007 22:33:15.970264 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_3d69b703-1e74-4e92-943d-6c73f029020e/watcher-api/0.log" Oct 07 22:33:23 crc kubenswrapper[4926]: I1007 22:33:23.679406 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:33:23 crc kubenswrapper[4926]: E1007 22:33:23.680084 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.305781 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fqt6b"] Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.308879 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.338886 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqt6b"] Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.504027 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-catalog-content\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.504417 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spqr9\" (UniqueName: \"kubernetes.io/projected/4a0a95bd-282a-4ba3-b56b-577f22284dee-kube-api-access-spqr9\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.504563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-utilities\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.606689 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-catalog-content\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.607180 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spqr9\" (UniqueName: \"kubernetes.io/projected/4a0a95bd-282a-4ba3-b56b-577f22284dee-kube-api-access-spqr9\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.607313 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-utilities\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.607644 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-utilities\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.607123 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-catalog-content\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.633809 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spqr9\" (UniqueName: \"kubernetes.io/projected/4a0a95bd-282a-4ba3-b56b-577f22284dee-kube-api-access-spqr9\") pod \"certified-operators-fqt6b\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.645177 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.680869 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:33:37 crc kubenswrapper[4926]: E1007 22:33:37.681066 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.909601 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-45fw5"] Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.911999 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:37 crc kubenswrapper[4926]: I1007 22:33:37.937326 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-45fw5"] Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.022766 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qqn4\" (UniqueName: \"kubernetes.io/projected/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-kube-api-access-6qqn4\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.022819 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-utilities\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.022999 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-catalog-content\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.124468 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qqn4\" (UniqueName: \"kubernetes.io/projected/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-kube-api-access-6qqn4\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.124754 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-utilities\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.124979 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-catalog-content\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.125247 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-utilities\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.125496 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-catalog-content\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.142158 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qqn4\" (UniqueName: \"kubernetes.io/projected/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-kube-api-access-6qqn4\") pod \"community-operators-45fw5\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.224147 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqt6b"] Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.249561 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.712595 4926 generic.go:334] "Generic (PLEG): container finished" podID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerID="8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8" exitCode=0 Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.712668 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerDied","Data":"8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8"} Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.712912 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerStarted","Data":"88aed71761222c8c6103267a4c2d9c07b4640978feca35bd1791af10dd5d9a1f"} Oct 07 22:33:38 crc kubenswrapper[4926]: W1007 22:33:38.735442 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb72a4c25_3bb3_435e_8af6_aef8968d0ad0.slice/crio-383d17c4029af0fdf33fbbbfced08755477230b43c70f0e7441fe4b269f78ca0 WatchSource:0}: Error finding container 383d17c4029af0fdf33fbbbfced08755477230b43c70f0e7441fe4b269f78ca0: Status 404 returned error can't find the container with id 383d17c4029af0fdf33fbbbfced08755477230b43c70f0e7441fe4b269f78ca0 Oct 07 22:33:38 crc kubenswrapper[4926]: I1007 22:33:38.738582 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-45fw5"] Oct 07 22:33:39 crc kubenswrapper[4926]: I1007 22:33:39.727209 4926 generic.go:334] "Generic (PLEG): container finished" podID="9764a0e8-f3f4-4206-9dc9-14a55ebf4344" containerID="90f32ab8e60f2834c1abfb62143c6dfb3d27313bbba119ae4e2e5d77dd1f7360" exitCode=0 Oct 07 22:33:39 crc kubenswrapper[4926]: I1007 22:33:39.727300 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" event={"ID":"9764a0e8-f3f4-4206-9dc9-14a55ebf4344","Type":"ContainerDied","Data":"90f32ab8e60f2834c1abfb62143c6dfb3d27313bbba119ae4e2e5d77dd1f7360"} Oct 07 22:33:39 crc kubenswrapper[4926]: I1007 22:33:39.730132 4926 generic.go:334] "Generic (PLEG): container finished" podID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerID="386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec" exitCode=0 Oct 07 22:33:39 crc kubenswrapper[4926]: I1007 22:33:39.730215 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerDied","Data":"386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec"} Oct 07 22:33:39 crc kubenswrapper[4926]: I1007 22:33:39.730254 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerStarted","Data":"383d17c4029af0fdf33fbbbfced08755477230b43c70f0e7441fe4b269f78ca0"} Oct 07 22:33:39 crc kubenswrapper[4926]: I1007 22:33:39.732043 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerStarted","Data":"230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7"} Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.747993 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerStarted","Data":"ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22"} Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.751566 4926 generic.go:334] "Generic (PLEG): container finished" podID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerID="230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7" exitCode=0 Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.751693 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerDied","Data":"230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7"} Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.892428 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.897877 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-host\") pod \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.898027 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9hsj\" (UniqueName: \"kubernetes.io/projected/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-kube-api-access-b9hsj\") pod \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\" (UID: \"9764a0e8-f3f4-4206-9dc9-14a55ebf4344\") " Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.899266 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-host" (OuterVolumeSpecName: "host") pod "9764a0e8-f3f4-4206-9dc9-14a55ebf4344" (UID: "9764a0e8-f3f4-4206-9dc9-14a55ebf4344"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.908453 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-kube-api-access-b9hsj" (OuterVolumeSpecName: "kube-api-access-b9hsj") pod "9764a0e8-f3f4-4206-9dc9-14a55ebf4344" (UID: "9764a0e8-f3f4-4206-9dc9-14a55ebf4344"). InnerVolumeSpecName "kube-api-access-b9hsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.971332 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-dt5qs"] Oct 07 22:33:40 crc kubenswrapper[4926]: I1007 22:33:40.979999 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-dt5qs"] Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.000678 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9hsj\" (UniqueName: \"kubernetes.io/projected/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-kube-api-access-b9hsj\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.000707 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9764a0e8-f3f4-4206-9dc9-14a55ebf4344-host\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.767790 4926 generic.go:334] "Generic (PLEG): container finished" podID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerID="ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22" exitCode=0 Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.767923 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerDied","Data":"ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22"} Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.773384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerStarted","Data":"ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03"} Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.777082 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54ce50dfe273ca439a3ab371b87398b5d701fc4459b954613c8ec8e2b1f0f981" Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.777179 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-dt5qs" Oct 07 22:33:41 crc kubenswrapper[4926]: I1007 22:33:41.823794 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fqt6b" podStartSLOduration=2.204841471 podStartE2EDuration="4.823773421s" podCreationTimestamp="2025-10-07 22:33:37 +0000 UTC" firstStartedPulling="2025-10-07 22:33:38.714340324 +0000 UTC m=+5868.752921474" lastFinishedPulling="2025-10-07 22:33:41.333272234 +0000 UTC m=+5871.371853424" observedRunningTime="2025-10-07 22:33:41.819685973 +0000 UTC m=+5871.858267163" watchObservedRunningTime="2025-10-07 22:33:41.823773421 +0000 UTC m=+5871.862354571" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.340206 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-695sl"] Oct 07 22:33:42 crc kubenswrapper[4926]: E1007 22:33:42.341276 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9764a0e8-f3f4-4206-9dc9-14a55ebf4344" containerName="container-00" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.341297 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9764a0e8-f3f4-4206-9dc9-14a55ebf4344" containerName="container-00" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.341572 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9764a0e8-f3f4-4206-9dc9-14a55ebf4344" containerName="container-00" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.342578 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.344951 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6c9g9"/"default-dockercfg-tccpd" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.433536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d13d082-e3f1-4c84-b393-3680859fd8a7-host\") pod \"crc-debug-695sl\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.433609 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8h6w\" (UniqueName: \"kubernetes.io/projected/6d13d082-e3f1-4c84-b393-3680859fd8a7-kube-api-access-b8h6w\") pod \"crc-debug-695sl\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.535379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d13d082-e3f1-4c84-b393-3680859fd8a7-host\") pod \"crc-debug-695sl\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.535480 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d13d082-e3f1-4c84-b393-3680859fd8a7-host\") pod \"crc-debug-695sl\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.535553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8h6w\" (UniqueName: \"kubernetes.io/projected/6d13d082-e3f1-4c84-b393-3680859fd8a7-kube-api-access-b8h6w\") pod \"crc-debug-695sl\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.566055 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8h6w\" (UniqueName: \"kubernetes.io/projected/6d13d082-e3f1-4c84-b393-3680859fd8a7-kube-api-access-b8h6w\") pod \"crc-debug-695sl\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.663913 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.702359 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9764a0e8-f3f4-4206-9dc9-14a55ebf4344" path="/var/lib/kubelet/pods/9764a0e8-f3f4-4206-9dc9-14a55ebf4344/volumes" Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.793181 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerStarted","Data":"fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead"} Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.794745 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-695sl" event={"ID":"6d13d082-e3f1-4c84-b393-3680859fd8a7","Type":"ContainerStarted","Data":"ea5ca69473cb3bb5684807384f318fe7f634c279e42028d03d5216c25e27b72d"} Oct 07 22:33:42 crc kubenswrapper[4926]: I1007 22:33:42.828460 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-45fw5" podStartSLOduration=3.381154983 podStartE2EDuration="5.828438577s" podCreationTimestamp="2025-10-07 22:33:37 +0000 UTC" firstStartedPulling="2025-10-07 22:33:39.733647184 +0000 UTC m=+5869.772228374" lastFinishedPulling="2025-10-07 22:33:42.180930818 +0000 UTC m=+5872.219511968" observedRunningTime="2025-10-07 22:33:42.820371984 +0000 UTC m=+5872.858953144" watchObservedRunningTime="2025-10-07 22:33:42.828438577 +0000 UTC m=+5872.867019747" Oct 07 22:33:43 crc kubenswrapper[4926]: I1007 22:33:43.809660 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d13d082-e3f1-4c84-b393-3680859fd8a7" containerID="b462dffc04e6dc92155749a0b50ec91b0a9b25393c4799252d661944b0bbe4dd" exitCode=0 Oct 07 22:33:43 crc kubenswrapper[4926]: I1007 22:33:43.809743 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-695sl" event={"ID":"6d13d082-e3f1-4c84-b393-3680859fd8a7","Type":"ContainerDied","Data":"b462dffc04e6dc92155749a0b50ec91b0a9b25393c4799252d661944b0bbe4dd"} Oct 07 22:33:43 crc kubenswrapper[4926]: E1007 22:33:43.876770 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d13d082_e3f1_4c84_b393_3680859fd8a7.slice/crio-conmon-b462dffc04e6dc92155749a0b50ec91b0a9b25393c4799252d661944b0bbe4dd.scope\": RecentStats: unable to find data in memory cache]" Oct 07 22:33:44 crc kubenswrapper[4926]: I1007 22:33:44.936523 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:44 crc kubenswrapper[4926]: I1007 22:33:44.987672 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d13d082-e3f1-4c84-b393-3680859fd8a7-host\") pod \"6d13d082-e3f1-4c84-b393-3680859fd8a7\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " Oct 07 22:33:44 crc kubenswrapper[4926]: I1007 22:33:44.987781 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d13d082-e3f1-4c84-b393-3680859fd8a7-host" (OuterVolumeSpecName: "host") pod "6d13d082-e3f1-4c84-b393-3680859fd8a7" (UID: "6d13d082-e3f1-4c84-b393-3680859fd8a7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:44 crc kubenswrapper[4926]: I1007 22:33:44.987844 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8h6w\" (UniqueName: \"kubernetes.io/projected/6d13d082-e3f1-4c84-b393-3680859fd8a7-kube-api-access-b8h6w\") pod \"6d13d082-e3f1-4c84-b393-3680859fd8a7\" (UID: \"6d13d082-e3f1-4c84-b393-3680859fd8a7\") " Oct 07 22:33:44 crc kubenswrapper[4926]: I1007 22:33:44.988322 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d13d082-e3f1-4c84-b393-3680859fd8a7-host\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:44 crc kubenswrapper[4926]: I1007 22:33:44.993647 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d13d082-e3f1-4c84-b393-3680859fd8a7-kube-api-access-b8h6w" (OuterVolumeSpecName: "kube-api-access-b8h6w") pod "6d13d082-e3f1-4c84-b393-3680859fd8a7" (UID: "6d13d082-e3f1-4c84-b393-3680859fd8a7"). InnerVolumeSpecName "kube-api-access-b8h6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:45 crc kubenswrapper[4926]: I1007 22:33:45.089246 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8h6w\" (UniqueName: \"kubernetes.io/projected/6d13d082-e3f1-4c84-b393-3680859fd8a7-kube-api-access-b8h6w\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:45 crc kubenswrapper[4926]: I1007 22:33:45.837245 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-695sl" event={"ID":"6d13d082-e3f1-4c84-b393-3680859fd8a7","Type":"ContainerDied","Data":"ea5ca69473cb3bb5684807384f318fe7f634c279e42028d03d5216c25e27b72d"} Oct 07 22:33:45 crc kubenswrapper[4926]: I1007 22:33:45.837289 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea5ca69473cb3bb5684807384f318fe7f634c279e42028d03d5216c25e27b72d" Oct 07 22:33:45 crc kubenswrapper[4926]: I1007 22:33:45.837337 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-695sl" Oct 07 22:33:47 crc kubenswrapper[4926]: I1007 22:33:47.645372 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:47 crc kubenswrapper[4926]: I1007 22:33:47.645718 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:47 crc kubenswrapper[4926]: I1007 22:33:47.690856 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:47 crc kubenswrapper[4926]: I1007 22:33:47.899024 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:47 crc kubenswrapper[4926]: I1007 22:33:47.945842 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqt6b"] Oct 07 22:33:48 crc kubenswrapper[4926]: I1007 22:33:48.249823 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:48 crc kubenswrapper[4926]: I1007 22:33:48.249865 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:48 crc kubenswrapper[4926]: I1007 22:33:48.306561 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:48 crc kubenswrapper[4926]: I1007 22:33:48.679026 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:33:48 crc kubenswrapper[4926]: E1007 22:33:48.679276 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:33:48 crc kubenswrapper[4926]: I1007 22:33:48.916725 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:49 crc kubenswrapper[4926]: I1007 22:33:49.874187 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fqt6b" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="registry-server" containerID="cri-o://ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03" gracePeriod=2 Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.330171 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-45fw5"] Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.336725 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.394500 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-utilities\") pod \"4a0a95bd-282a-4ba3-b56b-577f22284dee\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.394685 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spqr9\" (UniqueName: \"kubernetes.io/projected/4a0a95bd-282a-4ba3-b56b-577f22284dee-kube-api-access-spqr9\") pod \"4a0a95bd-282a-4ba3-b56b-577f22284dee\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.394739 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-catalog-content\") pod \"4a0a95bd-282a-4ba3-b56b-577f22284dee\" (UID: \"4a0a95bd-282a-4ba3-b56b-577f22284dee\") " Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.396457 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-utilities" (OuterVolumeSpecName: "utilities") pod "4a0a95bd-282a-4ba3-b56b-577f22284dee" (UID: "4a0a95bd-282a-4ba3-b56b-577f22284dee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.409383 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a0a95bd-282a-4ba3-b56b-577f22284dee-kube-api-access-spqr9" (OuterVolumeSpecName: "kube-api-access-spqr9") pod "4a0a95bd-282a-4ba3-b56b-577f22284dee" (UID: "4a0a95bd-282a-4ba3-b56b-577f22284dee"). InnerVolumeSpecName "kube-api-access-spqr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.496581 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spqr9\" (UniqueName: \"kubernetes.io/projected/4a0a95bd-282a-4ba3-b56b-577f22284dee-kube-api-access-spqr9\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.496608 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.526360 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a0a95bd-282a-4ba3-b56b-577f22284dee" (UID: "4a0a95bd-282a-4ba3-b56b-577f22284dee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.598083 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0a95bd-282a-4ba3-b56b-577f22284dee-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.884065 4926 generic.go:334] "Generic (PLEG): container finished" podID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerID="ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03" exitCode=0 Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.884393 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-45fw5" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="registry-server" containerID="cri-o://fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead" gracePeriod=2 Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.884724 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqt6b" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.885601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerDied","Data":"ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03"} Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.885636 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqt6b" event={"ID":"4a0a95bd-282a-4ba3-b56b-577f22284dee","Type":"ContainerDied","Data":"88aed71761222c8c6103267a4c2d9c07b4640978feca35bd1791af10dd5d9a1f"} Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.885656 4926 scope.go:117] "RemoveContainer" containerID="ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.919304 4926 scope.go:117] "RemoveContainer" containerID="230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7" Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.919332 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqt6b"] Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.930018 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fqt6b"] Oct 07 22:33:50 crc kubenswrapper[4926]: I1007 22:33:50.937322 4926 scope.go:117] "RemoveContainer" containerID="8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.070659 4926 scope.go:117] "RemoveContainer" containerID="ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03" Oct 07 22:33:51 crc kubenswrapper[4926]: E1007 22:33:51.076724 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03\": container with ID starting with ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03 not found: ID does not exist" containerID="ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.076773 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03"} err="failed to get container status \"ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03\": rpc error: code = NotFound desc = could not find container \"ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03\": container with ID starting with ff57d4c8a49947b6fc11c6857d6362175f6e3f77042b1edeb75f7ed01eda7d03 not found: ID does not exist" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.076802 4926 scope.go:117] "RemoveContainer" containerID="230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7" Oct 07 22:33:51 crc kubenswrapper[4926]: E1007 22:33:51.077270 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7\": container with ID starting with 230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7 not found: ID does not exist" containerID="230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.077313 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7"} err="failed to get container status \"230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7\": rpc error: code = NotFound desc = could not find container \"230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7\": container with ID starting with 230145c13f3ccc7fa7819068bc48594093af73212ebde52d1ab63459f6a168d7 not found: ID does not exist" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.077352 4926 scope.go:117] "RemoveContainer" containerID="8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8" Oct 07 22:33:51 crc kubenswrapper[4926]: E1007 22:33:51.077807 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8\": container with ID starting with 8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8 not found: ID does not exist" containerID="8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.077833 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8"} err="failed to get container status \"8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8\": rpc error: code = NotFound desc = could not find container \"8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8\": container with ID starting with 8f3e6f82c0fbd7064f90ac66a4428b79ea675818f10b76a4a378cef7537225a8 not found: ID does not exist" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.319144 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.415930 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qqn4\" (UniqueName: \"kubernetes.io/projected/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-kube-api-access-6qqn4\") pod \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.416042 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-utilities\") pod \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.416111 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-catalog-content\") pod \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\" (UID: \"b72a4c25-3bb3-435e-8af6-aef8968d0ad0\") " Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.417430 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-utilities" (OuterVolumeSpecName: "utilities") pod "b72a4c25-3bb3-435e-8af6-aef8968d0ad0" (UID: "b72a4c25-3bb3-435e-8af6-aef8968d0ad0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.422881 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-kube-api-access-6qqn4" (OuterVolumeSpecName: "kube-api-access-6qqn4") pod "b72a4c25-3bb3-435e-8af6-aef8968d0ad0" (UID: "b72a4c25-3bb3-435e-8af6-aef8968d0ad0"). InnerVolumeSpecName "kube-api-access-6qqn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.517876 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qqn4\" (UniqueName: \"kubernetes.io/projected/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-kube-api-access-6qqn4\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.518265 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.728294 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b72a4c25-3bb3-435e-8af6-aef8968d0ad0" (UID: "b72a4c25-3bb3-435e-8af6-aef8968d0ad0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.821883 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72a4c25-3bb3-435e-8af6-aef8968d0ad0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.896072 4926 generic.go:334] "Generic (PLEG): container finished" podID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerID="fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead" exitCode=0 Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.896225 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45fw5" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.896375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerDied","Data":"fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead"} Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.896405 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45fw5" event={"ID":"b72a4c25-3bb3-435e-8af6-aef8968d0ad0","Type":"ContainerDied","Data":"383d17c4029af0fdf33fbbbfced08755477230b43c70f0e7441fe4b269f78ca0"} Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.896423 4926 scope.go:117] "RemoveContainer" containerID="fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.928165 4926 scope.go:117] "RemoveContainer" containerID="ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.952786 4926 scope.go:117] "RemoveContainer" containerID="386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.974781 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-45fw5"] Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.974937 4926 scope.go:117] "RemoveContainer" containerID="fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead" Oct 07 22:33:51 crc kubenswrapper[4926]: E1007 22:33:51.975417 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead\": container with ID starting with fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead not found: ID does not exist" containerID="fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.975458 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead"} err="failed to get container status \"fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead\": rpc error: code = NotFound desc = could not find container \"fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead\": container with ID starting with fdaadb8531d7ef5c9cdb94b3e660862f485b9e18490d075c4b50aff4d3482ead not found: ID does not exist" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.975480 4926 scope.go:117] "RemoveContainer" containerID="ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22" Oct 07 22:33:51 crc kubenswrapper[4926]: E1007 22:33:51.976052 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22\": container with ID starting with ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22 not found: ID does not exist" containerID="ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.976071 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22"} err="failed to get container status \"ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22\": rpc error: code = NotFound desc = could not find container \"ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22\": container with ID starting with ce98b205e0274b33074941872d3ac1a00b69be695cd59991f3a6ac3c83edaf22 not found: ID does not exist" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.976085 4926 scope.go:117] "RemoveContainer" containerID="386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec" Oct 07 22:33:51 crc kubenswrapper[4926]: E1007 22:33:51.976438 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec\": container with ID starting with 386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec not found: ID does not exist" containerID="386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.976458 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec"} err="failed to get container status \"386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec\": rpc error: code = NotFound desc = could not find container \"386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec\": container with ID starting with 386fb06ab02633fecf9130e3d678eb77b2e01f9223805574d6e94fa7417ae7ec not found: ID does not exist" Oct 07 22:33:51 crc kubenswrapper[4926]: I1007 22:33:51.991118 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-45fw5"] Oct 07 22:33:52 crc kubenswrapper[4926]: I1007 22:33:52.688549 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" path="/var/lib/kubelet/pods/4a0a95bd-282a-4ba3-b56b-577f22284dee/volumes" Oct 07 22:33:52 crc kubenswrapper[4926]: I1007 22:33:52.689463 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" path="/var/lib/kubelet/pods/b72a4c25-3bb3-435e-8af6-aef8968d0ad0/volumes" Oct 07 22:33:52 crc kubenswrapper[4926]: I1007 22:33:52.904133 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-695sl"] Oct 07 22:33:52 crc kubenswrapper[4926]: I1007 22:33:52.910318 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-695sl"] Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.163528 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-n2sp6"] Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164421 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="registry-server" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164439 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="registry-server" Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164457 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="extract-content" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164465 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="extract-content" Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164494 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="extract-utilities" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164503 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="extract-utilities" Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164520 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="registry-server" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164527 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="registry-server" Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164545 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="extract-content" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164553 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="extract-content" Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164581 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="extract-utilities" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164589 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="extract-utilities" Oct 07 22:33:54 crc kubenswrapper[4926]: E1007 22:33:54.164610 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d13d082-e3f1-4c84-b393-3680859fd8a7" containerName="container-00" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164618 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d13d082-e3f1-4c84-b393-3680859fd8a7" containerName="container-00" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164845 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d13d082-e3f1-4c84-b393-3680859fd8a7" containerName="container-00" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164869 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72a4c25-3bb3-435e-8af6-aef8968d0ad0" containerName="registry-server" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.164892 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a0a95bd-282a-4ba3-b56b-577f22284dee" containerName="registry-server" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.165951 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.168856 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6c9g9"/"default-dockercfg-tccpd" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.270678 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9wlk\" (UniqueName: \"kubernetes.io/projected/5b08bd14-fb97-4405-8f3c-ff37609139d9-kube-api-access-k9wlk\") pod \"crc-debug-n2sp6\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.270826 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b08bd14-fb97-4405-8f3c-ff37609139d9-host\") pod \"crc-debug-n2sp6\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.373853 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b08bd14-fb97-4405-8f3c-ff37609139d9-host\") pod \"crc-debug-n2sp6\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.374000 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b08bd14-fb97-4405-8f3c-ff37609139d9-host\") pod \"crc-debug-n2sp6\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.374058 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9wlk\" (UniqueName: \"kubernetes.io/projected/5b08bd14-fb97-4405-8f3c-ff37609139d9-kube-api-access-k9wlk\") pod \"crc-debug-n2sp6\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.403378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9wlk\" (UniqueName: \"kubernetes.io/projected/5b08bd14-fb97-4405-8f3c-ff37609139d9-kube-api-access-k9wlk\") pod \"crc-debug-n2sp6\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.497906 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:54 crc kubenswrapper[4926]: W1007 22:33:54.543751 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b08bd14_fb97_4405_8f3c_ff37609139d9.slice/crio-f27f523d713af6946e03a0c3d11ac1a80dc6aa99b2c95d3acd3d9c7a337d2443 WatchSource:0}: Error finding container f27f523d713af6946e03a0c3d11ac1a80dc6aa99b2c95d3acd3d9c7a337d2443: Status 404 returned error can't find the container with id f27f523d713af6946e03a0c3d11ac1a80dc6aa99b2c95d3acd3d9c7a337d2443 Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.690948 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d13d082-e3f1-4c84-b393-3680859fd8a7" path="/var/lib/kubelet/pods/6d13d082-e3f1-4c84-b393-3680859fd8a7/volumes" Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.936985 4926 generic.go:334] "Generic (PLEG): container finished" podID="5b08bd14-fb97-4405-8f3c-ff37609139d9" containerID="7d6a61a190cc9fffda18a5b16bb18fa0b490044556e9f001727169421e888f01" exitCode=0 Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.937045 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" event={"ID":"5b08bd14-fb97-4405-8f3c-ff37609139d9","Type":"ContainerDied","Data":"7d6a61a190cc9fffda18a5b16bb18fa0b490044556e9f001727169421e888f01"} Oct 07 22:33:54 crc kubenswrapper[4926]: I1007 22:33:54.937076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" event={"ID":"5b08bd14-fb97-4405-8f3c-ff37609139d9","Type":"ContainerStarted","Data":"f27f523d713af6946e03a0c3d11ac1a80dc6aa99b2c95d3acd3d9c7a337d2443"} Oct 07 22:33:55 crc kubenswrapper[4926]: I1007 22:33:55.009548 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-n2sp6"] Oct 07 22:33:55 crc kubenswrapper[4926]: I1007 22:33:55.026876 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6c9g9/crc-debug-n2sp6"] Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.061081 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.121439 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b08bd14-fb97-4405-8f3c-ff37609139d9-host\") pod \"5b08bd14-fb97-4405-8f3c-ff37609139d9\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.121536 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9wlk\" (UniqueName: \"kubernetes.io/projected/5b08bd14-fb97-4405-8f3c-ff37609139d9-kube-api-access-k9wlk\") pod \"5b08bd14-fb97-4405-8f3c-ff37609139d9\" (UID: \"5b08bd14-fb97-4405-8f3c-ff37609139d9\") " Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.121577 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b08bd14-fb97-4405-8f3c-ff37609139d9-host" (OuterVolumeSpecName: "host") pod "5b08bd14-fb97-4405-8f3c-ff37609139d9" (UID: "5b08bd14-fb97-4405-8f3c-ff37609139d9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.122015 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b08bd14-fb97-4405-8f3c-ff37609139d9-host\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.128101 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b08bd14-fb97-4405-8f3c-ff37609139d9-kube-api-access-k9wlk" (OuterVolumeSpecName: "kube-api-access-k9wlk") pod "5b08bd14-fb97-4405-8f3c-ff37609139d9" (UID: "5b08bd14-fb97-4405-8f3c-ff37609139d9"). InnerVolumeSpecName "kube-api-access-k9wlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.224003 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9wlk\" (UniqueName: \"kubernetes.io/projected/5b08bd14-fb97-4405-8f3c-ff37609139d9-kube-api-access-k9wlk\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.688767 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b08bd14-fb97-4405-8f3c-ff37609139d9" path="/var/lib/kubelet/pods/5b08bd14-fb97-4405-8f3c-ff37609139d9/volumes" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.711062 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/util/0.log" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.894793 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/pull/0.log" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.896309 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/util/0.log" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.899558 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/pull/0.log" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.954763 4926 scope.go:117] "RemoveContainer" containerID="7d6a61a190cc9fffda18a5b16bb18fa0b490044556e9f001727169421e888f01" Oct 07 22:33:56 crc kubenswrapper[4926]: I1007 22:33:56.954915 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/crc-debug-n2sp6" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.048982 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/pull/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.072776 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/util/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.074542 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62467cf2462e1e88bd5a8471333ffd775b5a0b6c934b4c6e238954d834kp6k2_2ab95c76-4619-40f4-a661-36f224cecca8/extract/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.272407 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-7ltns_fea578e9-bbed-446d-beb9-5fcab1884b60/kube-rbac-proxy/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.284304 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-7ltns_fea578e9-bbed-446d-beb9-5fcab1884b60/manager/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.321539 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-bljj9_313c922a-ffe2-4d92-a221-bde117d0f196/kube-rbac-proxy/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.606379 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-pwsh8_a2ab6887-9ac0-4beb-8da0-8cb27da2ff66/kube-rbac-proxy/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.626092 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-bljj9_313c922a-ffe2-4d92-a221-bde117d0f196/manager/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.665315 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-pwsh8_a2ab6887-9ac0-4beb-8da0-8cb27da2ff66/manager/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.808454 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-jnbmb_bd230ff4-3566-43dc-be73-46cfe038bd76/kube-rbac-proxy/0.log" Oct 07 22:33:57 crc kubenswrapper[4926]: I1007 22:33:57.901110 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-jnbmb_bd230ff4-3566-43dc-be73-46cfe038bd76/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.013992 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-mwlfv_fc46d600-5c99-4ce4-9beb-06376dbacd8a/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.019433 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-mwlfv_fc46d600-5c99-4ce4-9beb-06376dbacd8a/kube-rbac-proxy/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.066062 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-nm2mk_6cb125b4-a707-4506-80fc-76bc981f1560/kube-rbac-proxy/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.203521 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-nm2mk_6cb125b4-a707-4506-80fc-76bc981f1560/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.256424 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-cgkzf_5517a3d5-4f45-44db-9552-fcd8de7356bf/kube-rbac-proxy/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.417461 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-cgkzf_5517a3d5-4f45-44db-9552-fcd8de7356bf/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.439564 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-zz2jk_d1f7536d-1b34-4d39-9249-b9959aa2a00f/kube-rbac-proxy/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.473785 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-zz2jk_d1f7536d-1b34-4d39-9249-b9959aa2a00f/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.617580 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-wxrzg_cefed49a-2a06-49a6-917e-ff05d6457bca/kube-rbac-proxy/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.693047 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-wxrzg_cefed49a-2a06-49a6-917e-ff05d6457bca/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.818946 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-tmh94_bd7d6078-1c87-459e-b83e-c408537f82a1/manager/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.829060 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-tmh94_bd7d6078-1c87-459e-b83e-c408537f82a1/kube-rbac-proxy/0.log" Oct 07 22:33:58 crc kubenswrapper[4926]: I1007 22:33:58.900476 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-7spjf_09117b6f-d508-460f-8848-ad0ba4f7c1be/kube-rbac-proxy/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.073589 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-tjfcp_ce47a525-41f3-4beb-b9b3-28d85831fdda/kube-rbac-proxy/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.086525 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-7spjf_09117b6f-d508-460f-8848-ad0ba4f7c1be/manager/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.172262 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-tjfcp_ce47a525-41f3-4beb-b9b3-28d85831fdda/manager/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.258062 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4xz96_0d0347d8-c6bc-4bd5-b75a-28ba94906a6d/kube-rbac-proxy/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.323529 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4xz96_0d0347d8-c6bc-4bd5-b75a-28ba94906a6d/manager/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.434714 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-rxdc8_2d16f135-1997-4bb6-9cb3-dae655640a01/kube-rbac-proxy/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.479442 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-rxdc8_2d16f135-1997-4bb6-9cb3-dae655640a01/manager/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.551262 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc_11159753-a963-44c9-bb63-53340a200de7/kube-rbac-proxy/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.636567 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665csxrqc_11159753-a963-44c9-bb63-53340a200de7/manager/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.679923 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:33:59 crc kubenswrapper[4926]: E1007 22:33:59.680624 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.700866 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-9554d4fbd-2cvhj_78f2ab6d-0297-49e4-91ff-ae7c8251fb0e/kube-rbac-proxy/0.log" Oct 07 22:33:59 crc kubenswrapper[4926]: I1007 22:33:59.939941 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8578696b-282fd_37e8795d-6050-4b61-9282-f3af17a926c8/kube-rbac-proxy/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.055948 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8578696b-282fd_37e8795d-6050-4b61-9282-f3af17a926c8/operator/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.154277 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hbjzj_240049fd-6fcf-493a-933e-c64ebca105b3/registry-server/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.349162 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-tsklh_f490b56a-2c4e-49ea-93fc-f2d8fb9454fd/kube-rbac-proxy/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.385656 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-tsklh_f490b56a-2c4e-49ea-93fc-f2d8fb9454fd/manager/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.570460 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-47m5t_d039dde6-16e1-42dd-a19f-b564a3939837/kube-rbac-proxy/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.583486 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-47m5t_d039dde6-16e1-42dd-a19f-b564a3939837/manager/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.689263 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-hx4t5_fb36bed3-c7f6-43ad-b250-da42097b673e/operator/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.833570 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-59xh2_35b344ad-2ef7-48e3-8643-e93e974fb326/manager/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.873014 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-59xh2_35b344ad-2ef7-48e3-8643-e93e974fb326/kube-rbac-proxy/0.log" Oct 07 22:34:00 crc kubenswrapper[4926]: I1007 22:34:00.895029 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-9554d4fbd-2cvhj_78f2ab6d-0297-49e4-91ff-ae7c8251fb0e/manager/0.log" Oct 07 22:34:01 crc kubenswrapper[4926]: I1007 22:34:01.048526 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-w2bgf_659a6e11-31f5-4a95-9eb1-3f61d8891ace/kube-rbac-proxy/0.log" Oct 07 22:34:01 crc kubenswrapper[4926]: I1007 22:34:01.084999 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-rtwpd_f825a6cc-9414-44f5-98a2-df2c3f1e825c/kube-rbac-proxy/0.log" Oct 07 22:34:01 crc kubenswrapper[4926]: I1007 22:34:01.140922 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-rtwpd_f825a6cc-9414-44f5-98a2-df2c3f1e825c/manager/0.log" Oct 07 22:34:01 crc kubenswrapper[4926]: I1007 22:34:01.230556 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-w2bgf_659a6e11-31f5-4a95-9eb1-3f61d8891ace/manager/0.log" Oct 07 22:34:01 crc kubenswrapper[4926]: I1007 22:34:01.279074 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-77fdd5cc9b-q8mkp_45317f9d-c1aa-4f42-b52f-f910a7082d1c/kube-rbac-proxy/0.log" Oct 07 22:34:01 crc kubenswrapper[4926]: I1007 22:34:01.437569 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-77fdd5cc9b-q8mkp_45317f9d-c1aa-4f42-b52f-f910a7082d1c/manager/0.log" Oct 07 22:34:13 crc kubenswrapper[4926]: I1007 22:34:13.678624 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:34:13 crc kubenswrapper[4926]: E1007 22:34:13.679583 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:34:16 crc kubenswrapper[4926]: I1007 22:34:16.721519 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6mlnn_5026deaa-f35e-4c06-9c90-cff95d6599a2/control-plane-machine-set-operator/0.log" Oct 07 22:34:16 crc kubenswrapper[4926]: I1007 22:34:16.892928 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qzh7t_24f2e598-e97c-4583-bc67-cd8afcf6065e/kube-rbac-proxy/0.log" Oct 07 22:34:16 crc kubenswrapper[4926]: I1007 22:34:16.952729 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qzh7t_24f2e598-e97c-4583-bc67-cd8afcf6065e/machine-api-operator/0.log" Oct 07 22:34:28 crc kubenswrapper[4926]: I1007 22:34:28.678864 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:34:28 crc kubenswrapper[4926]: E1007 22:34:28.679677 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:34:30 crc kubenswrapper[4926]: I1007 22:34:30.089616 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-v56pb_742ac735-fc39-4ab6-a072-834db359b8d8/cert-manager-controller/0.log" Oct 07 22:34:30 crc kubenswrapper[4926]: I1007 22:34:30.260816 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-77j9j_3aa5ce8c-a725-4bd2-a504-887c768a64b6/cert-manager-cainjector/0.log" Oct 07 22:34:30 crc kubenswrapper[4926]: I1007 22:34:30.314584 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dd6ts_07446e4f-68a2-422e-b303-e3637d956a4c/cert-manager-webhook/0.log" Oct 07 22:34:41 crc kubenswrapper[4926]: I1007 22:34:41.678559 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:34:41 crc kubenswrapper[4926]: E1007 22:34:41.679657 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:34:42 crc kubenswrapper[4926]: I1007 22:34:42.020366 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-wdx8h_7b429d99-24bf-4fee-81fb-eb1bc76d9788/nmstate-console-plugin/0.log" Oct 07 22:34:42 crc kubenswrapper[4926]: I1007 22:34:42.198743 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lswbm_8c97c92f-f0b8-48ec-b5f7-6594bd06c948/nmstate-handler/0.log" Oct 07 22:34:42 crc kubenswrapper[4926]: I1007 22:34:42.233814 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-mczl9_8daf114a-72f8-44e9-ac14-166bdf8e73cb/kube-rbac-proxy/0.log" Oct 07 22:34:42 crc kubenswrapper[4926]: I1007 22:34:42.307567 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-mczl9_8daf114a-72f8-44e9-ac14-166bdf8e73cb/nmstate-metrics/0.log" Oct 07 22:34:42 crc kubenswrapper[4926]: I1007 22:34:42.396949 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-vqnbn_c6ce449c-47a2-4bf7-9556-c3b72d502494/nmstate-operator/0.log" Oct 07 22:34:42 crc kubenswrapper[4926]: I1007 22:34:42.506773 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-r4nvj_aca14e9c-b601-4092-b7a5-72a80506b27e/nmstate-webhook/0.log" Oct 07 22:34:54 crc kubenswrapper[4926]: I1007 22:34:54.678762 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:34:54 crc kubenswrapper[4926]: E1007 22:34:54.679470 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:34:56 crc kubenswrapper[4926]: I1007 22:34:56.875784 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xdk65_17a8cd66-733c-4d61-83db-d0e56464b905/kube-rbac-proxy/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.069943 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xdk65_17a8cd66-733c-4d61-83db-d0e56464b905/controller/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.164117 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.244812 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.245153 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.314681 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.334773 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.522689 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.540526 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.540532 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.595856 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.732396 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-frr-files/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.770107 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-reloader/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.773115 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/cp-metrics/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.797770 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/controller/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.993911 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/frr-metrics/0.log" Oct 07 22:34:57 crc kubenswrapper[4926]: I1007 22:34:57.997129 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/kube-rbac-proxy/0.log" Oct 07 22:34:58 crc kubenswrapper[4926]: I1007 22:34:58.028277 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/kube-rbac-proxy-frr/0.log" Oct 07 22:34:58 crc kubenswrapper[4926]: I1007 22:34:58.195375 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/reloader/0.log" Oct 07 22:34:58 crc kubenswrapper[4926]: I1007 22:34:58.232584 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-dh249_5c050d21-9d7e-4e41-9a8e-991580804642/frr-k8s-webhook-server/0.log" Oct 07 22:34:58 crc kubenswrapper[4926]: I1007 22:34:58.520680 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74978ff576-w6wnd_fe0c800e-5525-4e10-8bb1-43a79192b2e8/manager/0.log" Oct 07 22:34:58 crc kubenswrapper[4926]: I1007 22:34:58.585131 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5f56c487b-tdcp8_472d3ee6-7ecc-471c-842c-94afa5bda368/webhook-server/0.log" Oct 07 22:34:58 crc kubenswrapper[4926]: I1007 22:34:58.728360 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mg6kr_138368c1-eaa9-42ba-8b93-67bf6396ef26/kube-rbac-proxy/0.log" Oct 07 22:34:59 crc kubenswrapper[4926]: I1007 22:34:59.288100 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mg6kr_138368c1-eaa9-42ba-8b93-67bf6396ef26/speaker/0.log" Oct 07 22:34:59 crc kubenswrapper[4926]: I1007 22:34:59.538804 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-86cft_d5f9b420-a8a0-4262-b6aa-49af0dc0e48c/frr/0.log" Oct 07 22:35:07 crc kubenswrapper[4926]: I1007 22:35:07.679765 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:35:07 crc kubenswrapper[4926]: E1007 22:35:07.682272 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:35:12 crc kubenswrapper[4926]: I1007 22:35:12.584316 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/util/0.log" Oct 07 22:35:12 crc kubenswrapper[4926]: I1007 22:35:12.813518 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/pull/0.log" Oct 07 22:35:12 crc kubenswrapper[4926]: I1007 22:35:12.875019 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/pull/0.log" Oct 07 22:35:12 crc kubenswrapper[4926]: I1007 22:35:12.895832 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/util/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.073144 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/extract/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.082599 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/util/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.127176 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vb9gz_b5d6db52-5370-492c-8641-7fa8e5601d4a/pull/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.256033 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/util/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.456156 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/pull/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.485728 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/util/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.491147 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/pull/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.641902 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/util/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.736909 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/pull/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.750960 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcd8pf_b7515481-e886-419f-a65b-ca89f922ef2e/extract/0.log" Oct 07 22:35:13 crc kubenswrapper[4926]: I1007 22:35:13.878549 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-utilities/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.061659 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-content/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.089484 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-utilities/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.117004 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-content/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.267743 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-utilities/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.291614 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/extract-content/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.495645 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-utilities/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.698241 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-utilities/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.752386 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-content/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.775755 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-content/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.972322 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-content/0.log" Oct 07 22:35:14 crc kubenswrapper[4926]: I1007 22:35:14.972604 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/extract-utilities/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.018209 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rbmzv_261aeb65-f6ef-4b6c-8fcc-8a1ae8b38509/registry-server/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.195435 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/util/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.462136 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/pull/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.490209 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/util/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.494777 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/pull/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.748113 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/util/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.776922 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/pull/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.807852 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c6jvt9_288387f2-b7aa-41d7-a10c-e11e65e054f1/extract/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.825390 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5p9gr_e3e61957-dad4-4160-8916-672ce8467461/registry-server/0.log" Oct 07 22:35:15 crc kubenswrapper[4926]: I1007 22:35:15.939623 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rt2mf_1ffbcf7c-a254-485e-8da4-69c2892ca0e9/marketplace-operator/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.015100 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-utilities/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.158091 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-content/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.167257 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-utilities/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.189281 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-content/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.320045 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-utilities/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.340701 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/extract-content/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.420805 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-utilities/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.551964 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncl5q_319f1329-c2a8-40eb-a8aa-bf7c4347fe47/registry-server/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.583483 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-content/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.606781 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-utilities/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.618326 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-content/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.801284 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-content/0.log" Oct 07 22:35:16 crc kubenswrapper[4926]: I1007 22:35:16.801638 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/extract-utilities/0.log" Oct 07 22:35:17 crc kubenswrapper[4926]: I1007 22:35:17.573671 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dq9zz_eb0efddf-5b03-4146-8ecb-ee37576f129c/registry-server/0.log" Oct 07 22:35:19 crc kubenswrapper[4926]: I1007 22:35:19.680413 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:35:19 crc kubenswrapper[4926]: E1007 22:35:19.681264 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5cd6_openshift-machine-config-operator(f78a9d2f-35fd-4558-a134-5c8fe26b0a4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" Oct 07 22:35:29 crc kubenswrapper[4926]: I1007 22:35:29.551506 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-hg4l9_0847160c-b3dd-4539-a141-d263faa41fd8/prometheus-operator/0.log" Oct 07 22:35:29 crc kubenswrapper[4926]: I1007 22:35:29.654457 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8546d9c7b6-58nvs_9f92dd43-5e91-41a3-bd84-8de059b4e9ee/prometheus-operator-admission-webhook/0.log" Oct 07 22:35:29 crc kubenswrapper[4926]: I1007 22:35:29.702252 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8546d9c7b6-nsg98_22bad45c-639a-4d86-aa08-1dd5f9b8ec5d/prometheus-operator-admission-webhook/0.log" Oct 07 22:35:29 crc kubenswrapper[4926]: I1007 22:35:29.841940 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-9pr8c_4d83bdfa-2009-43f5-93b1-cdc8e568020d/operator/0.log" Oct 07 22:35:29 crc kubenswrapper[4926]: I1007 22:35:29.904306 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-9vvdj_b4f527c2-fe8d-4da9-8477-2f02507a6178/perses-operator/0.log" Oct 07 22:35:33 crc kubenswrapper[4926]: I1007 22:35:33.680095 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" Oct 07 22:35:33 crc kubenswrapper[4926]: I1007 22:35:33.987374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"0d75445c781ffa46816ccce06d7682bc60e82e0778f842b3128c016e05afaa01"} Oct 07 22:35:39 crc kubenswrapper[4926]: E1007 22:35:39.654097 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.58:49196->38.102.83.58:42335: write tcp 38.102.83.58:49196->38.102.83.58:42335: write: connection reset by peer Oct 07 22:35:50 crc kubenswrapper[4926]: E1007 22:35:50.600985 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.58:43206->38.102.83.58:42335: write tcp 38.102.83.58:43206->38.102.83.58:42335: write: broken pipe Oct 07 22:37:33 crc kubenswrapper[4926]: I1007 22:37:33.210248 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:37:33 crc kubenswrapper[4926]: I1007 22:37:33.210975 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:37:40 crc kubenswrapper[4926]: I1007 22:37:40.468037 4926 generic.go:334] "Generic (PLEG): container finished" podID="2243e5f5-ea2c-4456-be75-6f10706aaa1f" containerID="2a59cffc4f7644ce4fd07a44aecfefa5bde6b30836fa72b42f94ee5bc4172e15" exitCode=0 Oct 07 22:37:40 crc kubenswrapper[4926]: I1007 22:37:40.468133 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6c9g9/must-gather-sschm" event={"ID":"2243e5f5-ea2c-4456-be75-6f10706aaa1f","Type":"ContainerDied","Data":"2a59cffc4f7644ce4fd07a44aecfefa5bde6b30836fa72b42f94ee5bc4172e15"} Oct 07 22:37:40 crc kubenswrapper[4926]: I1007 22:37:40.469589 4926 scope.go:117] "RemoveContainer" containerID="2a59cffc4f7644ce4fd07a44aecfefa5bde6b30836fa72b42f94ee5bc4172e15" Oct 07 22:37:41 crc kubenswrapper[4926]: I1007 22:37:41.044700 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6c9g9_must-gather-sschm_2243e5f5-ea2c-4456-be75-6f10706aaa1f/gather/0.log" Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.111034 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6c9g9/must-gather-sschm"] Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.111644 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6c9g9/must-gather-sschm" podUID="2243e5f5-ea2c-4456-be75-6f10706aaa1f" containerName="copy" containerID="cri-o://562eeb5031d037ce8862de91b9c9e07b8d941d0bac3fe902032947593becdd95" gracePeriod=2 Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.128495 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6c9g9/must-gather-sschm"] Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.635601 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6c9g9_must-gather-sschm_2243e5f5-ea2c-4456-be75-6f10706aaa1f/copy/0.log" Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.636372 4926 generic.go:334] "Generic (PLEG): container finished" podID="2243e5f5-ea2c-4456-be75-6f10706aaa1f" containerID="562eeb5031d037ce8862de91b9c9e07b8d941d0bac3fe902032947593becdd95" exitCode=143 Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.636426 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b055124dc058e942fc4c189f8cd82ea4d2a74d48dd05d65840df09ae60c7281c" Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.661620 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6c9g9_must-gather-sschm_2243e5f5-ea2c-4456-be75-6f10706aaa1f/copy/0.log" Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.662178 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.801768 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5clmg\" (UniqueName: \"kubernetes.io/projected/2243e5f5-ea2c-4456-be75-6f10706aaa1f-kube-api-access-5clmg\") pod \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.801841 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output\") pod \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.809700 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2243e5f5-ea2c-4456-be75-6f10706aaa1f-kube-api-access-5clmg" (OuterVolumeSpecName: "kube-api-access-5clmg") pod "2243e5f5-ea2c-4456-be75-6f10706aaa1f" (UID: "2243e5f5-ea2c-4456-be75-6f10706aaa1f"). InnerVolumeSpecName "kube-api-access-5clmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:37:54 crc kubenswrapper[4926]: I1007 22:37:54.903833 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5clmg\" (UniqueName: \"kubernetes.io/projected/2243e5f5-ea2c-4456-be75-6f10706aaa1f-kube-api-access-5clmg\") on node \"crc\" DevicePath \"\"" Oct 07 22:37:55 crc kubenswrapper[4926]: I1007 22:37:55.005239 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2243e5f5-ea2c-4456-be75-6f10706aaa1f" (UID: "2243e5f5-ea2c-4456-be75-6f10706aaa1f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:37:55 crc kubenswrapper[4926]: I1007 22:37:55.005804 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output\") pod \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\" (UID: \"2243e5f5-ea2c-4456-be75-6f10706aaa1f\") " Oct 07 22:37:55 crc kubenswrapper[4926]: W1007 22:37:55.006014 4926 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/2243e5f5-ea2c-4456-be75-6f10706aaa1f/volumes/kubernetes.io~empty-dir/must-gather-output Oct 07 22:37:55 crc kubenswrapper[4926]: I1007 22:37:55.006059 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2243e5f5-ea2c-4456-be75-6f10706aaa1f" (UID: "2243e5f5-ea2c-4456-be75-6f10706aaa1f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:37:55 crc kubenswrapper[4926]: I1007 22:37:55.006622 4926 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2243e5f5-ea2c-4456-be75-6f10706aaa1f-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 07 22:37:55 crc kubenswrapper[4926]: I1007 22:37:55.649393 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6c9g9/must-gather-sschm" Oct 07 22:37:56 crc kubenswrapper[4926]: I1007 22:37:56.690743 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2243e5f5-ea2c-4456-be75-6f10706aaa1f" path="/var/lib/kubelet/pods/2243e5f5-ea2c-4456-be75-6f10706aaa1f/volumes" Oct 07 22:38:03 crc kubenswrapper[4926]: I1007 22:38:03.209761 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:38:03 crc kubenswrapper[4926]: I1007 22:38:03.210373 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:38:07 crc kubenswrapper[4926]: I1007 22:38:07.603928 4926 scope.go:117] "RemoveContainer" containerID="90f32ab8e60f2834c1abfb62143c6dfb3d27313bbba119ae4e2e5d77dd1f7360" Oct 07 22:38:07 crc kubenswrapper[4926]: I1007 22:38:07.626351 4926 scope.go:117] "RemoveContainer" containerID="2a59cffc4f7644ce4fd07a44aecfefa5bde6b30836fa72b42f94ee5bc4172e15" Oct 07 22:38:07 crc kubenswrapper[4926]: I1007 22:38:07.731522 4926 scope.go:117] "RemoveContainer" containerID="562eeb5031d037ce8862de91b9c9e07b8d941d0bac3fe902032947593becdd95" Oct 07 22:38:33 crc kubenswrapper[4926]: I1007 22:38:33.209748 4926 patch_prober.go:28] interesting pod/machine-config-daemon-t5cd6 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:38:33 crc kubenswrapper[4926]: I1007 22:38:33.210517 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:38:33 crc kubenswrapper[4926]: I1007 22:38:33.210590 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" Oct 07 22:38:33 crc kubenswrapper[4926]: I1007 22:38:33.211758 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d75445c781ffa46816ccce06d7682bc60e82e0778f842b3128c016e05afaa01"} pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:38:33 crc kubenswrapper[4926]: I1007 22:38:33.211866 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" podUID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerName="machine-config-daemon" containerID="cri-o://0d75445c781ffa46816ccce06d7682bc60e82e0778f842b3128c016e05afaa01" gracePeriod=600 Oct 07 22:38:34 crc kubenswrapper[4926]: I1007 22:38:34.090029 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78a9d2f-35fd-4558-a134-5c8fe26b0a4a" containerID="0d75445c781ffa46816ccce06d7682bc60e82e0778f842b3128c016e05afaa01" exitCode=0 Oct 07 22:38:34 crc kubenswrapper[4926]: I1007 22:38:34.090100 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerDied","Data":"0d75445c781ffa46816ccce06d7682bc60e82e0778f842b3128c016e05afaa01"} Oct 07 22:38:34 crc kubenswrapper[4926]: I1007 22:38:34.090761 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5cd6" event={"ID":"f78a9d2f-35fd-4558-a134-5c8fe26b0a4a","Type":"ContainerStarted","Data":"3223a9199e7aa5ecc5453d7bee29c9d9aaca861637c603037689e2f7b1782250"} Oct 07 22:38:34 crc kubenswrapper[4926]: I1007 22:38:34.090788 4926 scope.go:117] "RemoveContainer" containerID="ae45bbfe5a095731592040fd6639c42f624012db3f479dfe336884727b29a984" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071313234024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071313235017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071276556016523 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071276556015473 5ustar corecore